-
Notifications
You must be signed in to change notification settings - Fork 5.6k
/
Copy pathauto_archive.py
358 lines (298 loc) · 12.8 KB
/
auto_archive.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
# Copyright 2020 MongoDB Inc.
#
# Permission is hereby granted, free of charge, to any person obtaining
# a copy of this software and associated documentation files (the
# "Software"), to deal in the Software without restriction, including
# without limitation the rights to use, copy, modify, merge, publish,
# distribute, sublicense, and/or sell copies of the Software, and to
# permit persons to whom the Software is furnished to do so, subject to
# the following conditions:
#
# The above copyright notice and this permission notice shall be included
# in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY
# KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE
# WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
# LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
# OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
# WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
#
import os
import sys
import SCons
PACKAGE_ALIAS_MAP = "AIB_PACKAGE_ALIAS_MAP"
AUTO_ARCHIVE_MAKE_ARCHIVE_CONTENT = """
import os
import sys
USAGE = '''
Usage: {} ARCHIVE_TYPE ARCHIVE_NAME ROOT_DIRECTORY FILES...
FILES should be absolute paths or relative to ROOT_DIRECTORY.
ARCHIVE_TYPE is one of zip or tar.
'''
if __name__ == "__main__":
if len(sys.argv) < 4:
print(sys.argv[0], "takes at minimum four arguments.")
print(USAGE.format(sys.argv[0]))
sys.exit(1)
archive_type = sys.argv[1]
archive_name = sys.argv[2]
root_dir = sys.argv[3]
file_list = sys.argv[4]
files = []
with open(file_list) as f:
files = f.read().splitlines()
if archive_type not in ("zip", "tar"):
print("unsupported archive_type", archive_type)
print(USAGE.format(sys.argv[0]))
sys.exit(1)
if archive_type == "zip":
import zipfile
archive = zipfile.ZipFile(archive_name, mode='w', compression=zipfile.ZIP_DEFLATED)
add_file = archive.write
else:
import tarfile
archive = tarfile.open(archive_name, mode='w:gz')
add_file = archive.add
os.chdir(root_dir)
for filename in files:
add_file(filename)
archive.close()
"""
def add_package_name_alias(env, component, role, name):
"""Add a package name mapping for the combination of component and role."""
# Verify we didn't get a None or empty string for any argument
if not name:
raise Exception("when setting a package name alias must provide a name parameter")
if not component:
raise Exception("No component provided for package name alias")
if not role:
raise Exception("No role provided for package name alias")
env[PACKAGE_ALIAS_MAP][(component, role)] = name
def get_package_name(env, component, role):
"""Return the package file name for the component and role combination."""
basename = env[PACKAGE_ALIAS_MAP].get(
# TODO: silent roles shouldn't be included here
(component, role),
"{component}-{role}".format(component=component, role=role),
)
return basename
def collect_transitive_files(env, entry):
"""
Collect all installed and transitively installed files for entry.
"""
cache = set()
files = []
stack = [entry]
# Find all the files directly contained in the component DAG for entry and
# it's dependencies.
while stack:
s = stack.pop()
if s in cache:
continue
cache.add(s)
stack.extend(s.dependencies)
files.extend(s.files)
cache.clear()
files, stack = stack, files
# Now we will call the scanner to find the transtive files of any files that
# we found from the component DAG.
bazel_installed = set()
while stack:
s = stack.pop()
if s in cache:
continue
cache.add(s)
files.append(s)
# scan_for_transitive_install is memoized so it's safe to call it in
# this loop. If it hasn't already run for a file we need to run it
# anyway.
stack.extend(env.GetTransitivelyInstalledFiles(s))
# if the current file is a bazel target we need to find its bazel deps
# and add them to the archive.
if env.GetOption("ninja") == "disabled" and env.GetOption("link-model") == "dynamic":
env.BazelAutoArchive(s, bazel_installed, stack)
real_node = getattr(s.attributes, "AIB_INSTALL_FROM", s)
# usually you might use scons .children call but we know we only care about
# direct libdeps in this case as they may be transition nodes that are between
# scons and bazels graph.
for child in getattr(real_node.attributes, "libdeps_direct_sorted", []):
try:
bazel_libdep = env.File(f"#/{env['SCONS2BAZEL_TARGETS'].bazel_output(child)}")
install_file = env.GetAutoInstalledFiles(bazel_libdep)
if not install_file:
shlib_suffix = env.subst("$SHLIBSUFFIX")
env.BazelAutoInstall(bazel_libdep, shlib_suffix)
install_file = env.GetAutoInstalledFiles(bazel_libdep)
env.BazelAutoArchive(install_file[0], bazel_installed, stack)
except KeyError:
pass
# Setting the AIB_NO_ARCHIVE attribute to True prevents outputs from an
# AutoInstall builder from being included into archives produced by this
# tool
# Usage:
# node = env.AutoInstall(...)
# setattr(node[0].attributes, 'AIB_NO_ARCHIVE', True)
return sorted(f for f in files if not getattr(f.attributes, "AIB_NO_ARCHIVE", False))
def auto_archive_gen(first_env, make_archive_script, pkg_fmt):
"""Generate an archive task function for pkg_fmt where pkg_fmt is one of zip, tar, or auto."""
if pkg_fmt == "auto":
if first_env["PLATFORM"] == "win32":
pkg_fmt = "zip"
else:
pkg_fmt = "tar"
def auto_archive(env, component, role):
pkg_name = get_package_name(env, component, role)
install_alias = "install-{component}{role}".format(
component=component,
role="" if env.GetRoleDeclaration(role).silent else "-" + role,
)
if pkg_fmt == "zip":
pkg_suffix = "$AUTO_ARCHIVE_ZIP_SUFFIX"
else:
pkg_suffix = "$AUTO_ARCHIVE_TARBALL_SUFFIX"
archive = env.AutoArchive(
target="$PKGDIR/{}.{}".format(pkg_name, pkg_suffix),
source=[make_archive_script] + env.Alias(install_alias),
__AUTO_ARCHIVE_TYPE=pkg_fmt,
AIB_COMPONENT=component,
AIB_ROLE=role,
)
# TODO: perhaps caching of packages / tarballs should be
# configurable? It's possible someone would want to do it.
env.NoCache(archive)
return archive
return auto_archive
def archive_builder(source, target, env, for_signature):
"""Build archives of the AutoInstall'd sources."""
if not source:
return []
source = env.Flatten([source])
common_ancestor = None
# Get the path elements that make up both DESTDIR and PREFIX. Then
# iterate the dest_dir_elems with the prefix path elements
# stripped off the end of the path converting them to strings for
# joining to make the common_ancestor.
#
# We pass the common_ancestor to tar via -C so that $PREFIX is
# preserved in the tarball.
common_ancestor = env.Dir("$DESTDIR")
archive_type = env["__AUTO_ARCHIVE_TYPE"]
make_archive_script = source[0]
compression_flags = ""
tar_cmd = env.WhereIs("tar")
if archive_type == "tar" and tar_cmd:
pigz_cmd = env.WhereIs("pigz")
if pigz_cmd:
# pigz is the parallel implementation of gizp,
# it uses all available cores on the machine.
# if available we use it to speedup compression.
compression_flags = "--use-compress-program='{pigz_cmd}'".format(pigz_cmd=pigz_cmd)
else:
compression_flags = "-z"
command_prefix = (
"{tar} -vc {compression_flags} -C {common_ancestor} -T {file_list} -f {archive_name}"
)
else:
command_prefix = "{python} {make_archive_script} {archive_type} {archive_name} {common_ancestor} {file_list}"
archive_name = env.File(target[0])
command_sig = command_prefix.format(
tar=tar_cmd,
compression_flags=compression_flags,
python=sys.executable,
archive_type=archive_type,
archive_name=archive_name,
make_archive_script=make_archive_script,
common_ancestor=common_ancestor,
file_list="",
)
# If we are just being invoked for our signature, we can omit the indirect dependencies
# found by expanding the transitive dependencies, since we really only have a hard dependency
# on our direct dependencies.
if for_signature:
return command_sig
component = env["AIB_COMPONENT"]
role = env["AIB_ROLE"]
entry = env["AIB_ALIAS_MAP"][component][role]
# Pre-process what should be in the archive. We need to pass the
# set of known installed files along to the transitive dependency
# walk so we can filter out files that aren't in the install
# directory.
installed = set(env.FindInstalledFiles())
# Collect all the installed files for our entry. This is doing a pure DAG
# walk idea of what should be. So we filter out any that are not in the
# installed set.
transitive_files = [f for f in collect_transitive_files(env, entry) if f in installed]
if not transitive_files:
return []
# TODO: relpath is costly, and we do it for every file in the archive here.
# We should find a way to avoid the repeated relpath invocation, probably by
# bucketing by directory.
relative_files = [
os.path.relpath(file.get_abspath(), common_ancestor.get_abspath())
for file in transitive_files
]
# This is not great for ninja, essentially we are doing realtime operations here, which is
# not terrible for scons CommandActionGenerators, because the generation happens right before
# executing the command. However, this means for ninja that the realtime things happen during
# ninja generation, and are far removed from ninja execution. Even if we split this into a
# separate action for ninja's sake, there would still be issues because the reason to make
# such a filelist is to prevent creating command lines which are too long, and actions must
# be converted to command lines for ninja. When the ninja tool is able to process scons
# callbacks in order and not via aggregation then this could be moved to a simple Textfile call.
file_list = str(target[0].abspath) + ".filelist"
os.makedirs(os.path.dirname(file_list), exist_ok=True)
with open(file_list, "w") as f:
for file in relative_files:
f.write(file + "\n")
cmd = command_prefix.format(
tar=tar_cmd,
compression_flags=compression_flags,
python=sys.executable,
archive_type=archive_type,
archive_name=archive_name,
make_archive_script=make_archive_script,
common_ancestor=common_ancestor,
file_list=file_list,
)
if env.GetOption("ninja") != "disabled":
if env.TargetOSIs("windows"):
cmd = 'echo "archive not supported with ninja, use scons only." & exit /b 1'
else:
cmd = 'echo "archive not supported with ninja, use scons only."; exit 1'
return cmd
def exists(env):
return True
def generate(env):
if not env.get("AUTO_INSTALL_ENABLED"):
env.Tool("auto_install_binaries")
bld = SCons.Builder.Builder(
action=SCons.Action.CommandGeneratorAction(
archive_builder,
{"cmdstr": "Building package ${TARGETS[0]} from ${SOURCES[1:]}"}
if not env.Verbose()
else {"cmdstr": ""},
)
)
env.Append(BUILDERS={"AutoArchive": bld})
env["AUTO_ARCHIVE_TARBALL_SUFFIX"] = env.get(
"AUTO_ARCHIVE_TARBALL_SUFFIX",
"tar.gz",
)
env["AUTO_ARCHIVE_ZIP_SUFFIX"] = env.get("AUTO_ARCHIVE_ZIP_SUFFIX", "zip")
env[PACKAGE_ALIAS_MAP] = {}
env.AddMethod(add_package_name_alias, "AddPackageNameAlias")
# TODO: $BUILD_ROOT should be $VARIANT_DIR after we fix our dir
# setup later on.
make_archive_script = env.Textfile(
target="$BUILD_ROOT/aib_make_archive.py",
source=[AUTO_ARCHIVE_MAKE_ARCHIVE_CONTENT],
)
env.AppendUnique(
AIB_TASKS={
"tar": (auto_archive_gen(env, make_archive_script, "tar"), False),
"zip": (auto_archive_gen(env, make_archive_script, "zip"), False),
"archive": (auto_archive_gen(env, make_archive_script, "auto"), False),
}
)