forked from pytorch/test-infra
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathmanage.py
executable file
·651 lines (585 loc) · 22.3 KB
/
manage.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
#!/usr/bin/env python
import argparse
import base64
import concurrent.futures
import dataclasses
import functools
import time
from contextlib import suppress
from os import path, makedirs
from datetime import datetime
from collections import defaultdict
from typing import Iterable, List, Type, Dict, Set, TypeVar, Optional
from re import sub, match, search
from packaging.version import parse as _parse_version, Version, InvalidVersion
import boto3
import botocore
S3 = boto3.resource('s3')
CLIENT = boto3.client('s3')
# bucket for download.pytorch.org
BUCKET = S3.Bucket('pytorch')
# bucket mirror just to hold index used with META CDN
BUCKET_META_CDN = S3.Bucket('pytorch-test')
INDEX_BUCKETS = {BUCKET, BUCKET_META_CDN}
ACCEPTED_FILE_EXTENSIONS = ("whl", "zip", "tar.gz")
ACCEPTED_SUBDIR_PATTERNS = [
r"cu[0-9]+", # for cuda
r"rocm[0-9]+\.[0-9]+", # for rocm
"cpu",
"xpu",
]
PREFIXES = [
"whl",
"whl/nightly",
"whl/test",
"libtorch",
"libtorch/nightly",
]
# NOTE: This refers to the name on the wheels themselves and not the name of
# package as specified by setuptools, for packages with "-" (hyphens) in their
# names you need to convert them to "_" (underscores) in order for them to be
# allowed here since the name of the wheels is compared here
PACKAGE_ALLOW_LIST = {x.lower() for x in [
# ---- torchtune additional packages ----
"aiohttp",
"aiosignal",
"aiohappyeyeballs",
"antlr4_python3_runtime",
"antlr4-python3-runtime",
"async_timeout",
"attrs",
"blobfile",
"datasets",
"dill",
"frozenlist",
"huggingface_hub",
"llnl_hatchet",
"lxml",
"multidict",
"multiprocess",
"omegaconf",
"pandas",
"psutil",
"pyarrow",
"pyarrow_hotfix",
"pycryptodomex",
"python_dateutil",
"pytz",
"PyYAML",
"regex",
"safetensors",
"sentencepiece",
"six",
"tiktoken",
"torchao",
"torchao_nightly",
"tzdata",
"xxhash",
"yarl",
# ---- triton additional packages ----
"Arpeggio",
"caliper_reader",
"contourpy",
"cycler",
"dill",
"fonttools",
"kiwisolver",
"llnl-hatchet",
"matplotlib",
"pandas",
"pydot",
"pyparsing",
"pytz",
"textx",
"tzdata",
"importlib_metadata",
"importlib_resources",
"zipp",
# ---- torch xpu additional packages ----
"dpcpp_cpp_rt",
"intel_cmplr_lib_rt",
"intel_cmplr_lib_ur",
"intel_cmplr_lic_rt",
"intel_opencl_rt",
"intel_sycl_rt",
"intel_openmp",
"tcmlib",
"umf",
"intel_pti",
# ----
"Pillow",
"certifi",
"charset_normalizer",
"cmake",
"colorama",
"fbgemm_gpu",
"filelock",
"fsspec",
"idna",
"iopath",
"intel_openmp",
"Jinja2",
"lit",
"lightning_utilities",
"MarkupSafe",
"mpmath",
"mkl",
"mypy_extensions",
"nestedtensor",
"networkx",
"numpy",
"nvidia_cublas_cu11",
"nvidia_cuda_cupti_cu11",
"nvidia_cuda_nvrtc_cu11",
"nvidia_cuda_runtime_cu11",
"nvidia_cudnn_cu11",
"nvidia_cufft_cu11",
"nvidia_curand_cu11",
"nvidia_cusolver_cu11",
"nvidia_cusparse_cu11",
"nvidia_nccl_cu11",
"nvidia_nvtx_cu11",
"nvidia_cublas_cu12",
"nvidia_cuda_cupti_cu12",
"nvidia_cuda_nvrtc_cu12",
"nvidia_cuda_runtime_cu12",
"nvidia_cudnn_cu12",
"nvidia_cufft_cu12",
"nvidia_curand_cu12",
"nvidia_cusolver_cu12",
"nvidia_cusparse_cu12",
"nvidia_cusparselt_cu12",
"nvidia_nccl_cu12",
"nvidia_nvtx_cu12",
"nvidia_nvjitlink_cu12",
"packaging",
"portalocker",
"pyre_extensions",
"pytorch_triton",
"pytorch_triton_rocm",
"pytorch_triton_xpu",
"requests",
"sympy",
"tbb",
"torch_no_python",
"torch",
"torch_tensorrt",
"torcharrow",
"torchaudio",
"torchcodec",
"torchcsprng",
"torchdata",
"torchdistx",
"torchmetrics",
"torchrec",
"torchtext",
"torchtune",
"torchvision",
"triton",
"tqdm",
"typing_extensions",
"typing_inspect",
"urllib3",
"xformers",
"executorch",
"setuptools",
"wheel",
]}
# Should match torch-2.0.0.dev20221221+cu118-cp310-cp310-linux_x86_64.whl as:
# Group 1: torch-2.0.0.dev
# Group 2: 20221221
PACKAGE_DATE_REGEX = r"([a-zA-z]*-[0-9.]*.dev)([0-9]*)"
# How many packages should we keep of a specific package?
KEEP_THRESHOLD = 60
# TODO (huydhn): Clean this up once ExecuTorch has a new stable release that
# match PyTorch stable release cadence. This nightly version is currently
# referred to publicly in ExecuTorch alpha 0.1 release. So we want to keep
# nightly binaries around for now
KEEP_NIGHTLY_PACKAGES_FOR_EXECUTORCH = {datetime(2023, 10, 10, 0, 0)}
S3IndexType = TypeVar('S3IndexType', bound='S3Index')
@dataclasses.dataclass(frozen=False)
@functools.total_ordering
class S3Object:
key: str
orig_key: str
checksum: Optional[str]
size: Optional[int]
pep658: Optional[str]
def __hash__(self):
return hash(self.key)
def __str__(self):
return self.key
def __eq__(self, other):
return self.key == other.key
def __lt__(self, other):
return self.key < other.key
def extract_package_build_time(full_package_name: str) -> datetime:
result = search(PACKAGE_DATE_REGEX, full_package_name)
if result is not None:
with suppress(ValueError):
# Ignore any value errors since they probably shouldn't be hidden anyways
return datetime.strptime(result.group(2), "%Y%m%d")
return datetime.now()
def between_bad_dates(package_build_time: datetime):
start_bad = datetime(year=2022, month=8, day=17)
end_bad = datetime(year=2022, month=12, day=30)
return start_bad <= package_build_time <= end_bad
def safe_parse_version(ver_str: str) -> Version:
try:
return _parse_version(ver_str)
except InvalidVersion:
return Version("0.0.0")
class S3Index:
def __init__(self: S3IndexType, objects: List[S3Object], prefix: str) -> None:
self.objects = objects
self.prefix = prefix.rstrip("/")
self.html_name = "index.html"
# should dynamically grab subdirectories like whl/test/cu101
# so we don't need to add them manually anymore
self.subdirs = {
path.dirname(obj.key) for obj in objects if path.dirname != prefix
}
def nightly_packages_to_show(self: S3IndexType) -> List[S3Object]:
"""Finding packages to show based on a threshold we specify
Basically takes our S3 packages, normalizes the version for easier
comparisons, then iterates over normalized versions until we reach a
threshold and then starts adding package to delete after that threshold
has been reached
After figuring out what versions we'd like to hide we iterate over
our original object list again and pick out the full paths to the
packages that are included in the list of versions to delete
"""
# also includes versions without GPU specifier (i.e. cu102) for easier
# sorting, sorts in reverse to put the most recent versions first
all_sorted_packages = sorted(
{self.normalize_package_version(obj) for obj in self.objects},
key=lambda name_ver: safe_parse_version(name_ver.split('-', 1)[-1]),
reverse=True,
)
packages: Dict[str, int] = defaultdict(int)
to_hide: Set[str] = set()
for obj in all_sorted_packages:
full_package_name = path.basename(obj)
package_name = full_package_name.split('-')[0]
package_build_time = extract_package_build_time(full_package_name)
# Hard pass on packages that are included in our allow list
if package_name.lower() not in PACKAGE_ALLOW_LIST:
to_hide.add(obj)
continue
if package_build_time not in KEEP_NIGHTLY_PACKAGES_FOR_EXECUTORCH and (
packages[package_name] >= KEEP_THRESHOLD
or between_bad_dates(package_build_time)
):
to_hide.add(obj)
else:
packages[package_name] += 1
return list(set(self.objects).difference({
obj for obj in self.objects
if self.normalize_package_version(obj) in to_hide
}))
def is_obj_at_root(self, obj: S3Object) -> bool:
return path.dirname(obj.key) == self.prefix
def _resolve_subdir(self, subdir: Optional[str] = None) -> str:
if not subdir:
subdir = self.prefix
# make sure we strip any trailing slashes
return subdir.rstrip("/")
def gen_file_list(
self,
subdir: Optional[str] = None,
package_name: Optional[str] = None
) -> Iterable[S3Object]:
objects = self.objects
subdir = self._resolve_subdir(subdir) + '/'
for obj in objects:
if package_name is not None and self.obj_to_package_name(obj) != package_name:
continue
if self.is_obj_at_root(obj) or obj.key.startswith(subdir):
yield obj
def get_package_names(self, subdir: Optional[str] = None) -> List[str]:
return sorted({self.obj_to_package_name(obj) for obj in self.gen_file_list(subdir)})
def normalize_package_version(self: S3IndexType, obj: S3Object) -> str:
# removes the GPU specifier from the package name as well as
# unnecessary things like the file extension, architecture name, etc.
return sub(
r"%2B.*",
"",
"-".join(path.basename(obj.key).split("-")[:2])
)
def obj_to_package_name(self, obj: S3Object) -> str:
return path.basename(obj.key).split('-', 1)[0].lower()
def to_libtorch_html(
self,
subdir: Optional[str] = None
) -> str:
"""Generates a string that can be used as the HTML index
Takes our objects and transforms them into HTML that have historically
been used by pip for installing pytorch, but now only used to generate libtorch browseable folder.
"""
out: List[str] = []
subdir = self._resolve_subdir(subdir)
is_root = subdir == self.prefix
for obj in self.gen_file_list(subdir, "libtorch"):
# Skip root objs, as they are irrelevant for libtorch indexes
if not is_root and self.is_obj_at_root(obj):
continue
# Strip our prefix
sanitized_obj = obj.key.replace(subdir, "", 1)
if sanitized_obj.startswith('/'):
sanitized_obj = sanitized_obj.lstrip("/")
out.append(f'<a href="/{obj.key}">{sanitized_obj}</a><br/>')
return "\n".join(sorted(out))
def to_simple_package_html(
self,
subdir: Optional[str],
package_name: str
) -> str:
"""Generates a string that can be used as the package simple HTML index
"""
out: List[str] = []
# Adding html header
out.append('<!DOCTYPE html>')
out.append('<html>')
out.append(' <body>')
out.append(' <h1>Links for {}</h1>'.format(package_name.lower().replace("_", "-")))
for obj in sorted(self.gen_file_list(subdir, package_name)):
maybe_fragment = f"#sha256={obj.checksum}" if obj.checksum else ""
# Temporary skip assigning sha256 to nightly index
# to be reverted on Jan 24, 2025.
if subdir is not None and "nightly" in subdir:
maybe_fragment = ""
pep658_attribute = ""
if obj.pep658:
pep658_sha = f"sha256={obj.pep658}"
# pep714 renames the attribute to data-core-metadata
pep658_attribute = (
f' data-dist-info-metadata="{pep658_sha}" data-core-metadata="{pep658_sha}"'
)
out.append(
f' <a href="/{obj.key}{maybe_fragment}"{pep658_attribute}>{path.basename(obj.key).replace("%2B","+")}</a><br/>'
)
# Adding html footer
out.append(' </body>')
out.append('</html>')
out.append(f'<!--TIMESTAMP {int(time.time())}-->')
return '\n'.join(out)
def to_simple_packages_html(
self,
subdir: Optional[str],
) -> str:
"""Generates a string that can be used as the simple HTML index
"""
out: List[str] = []
# Adding html header
out.append('<!DOCTYPE html>')
out.append('<html>')
out.append(' <body>')
for pkg_name in sorted(self.get_package_names(subdir)):
out.append(f' <a href="{pkg_name.lower().replace("_","-")}/">{pkg_name.replace("_","-")}</a><br/>')
# Adding html footer
out.append(' </body>')
out.append('</html>')
out.append(f'<!--TIMESTAMP {int(time.time())}-->')
return '\n'.join(out)
def upload_libtorch_html(self) -> None:
for subdir in self.subdirs:
index_html = self.to_libtorch_html(subdir=subdir)
for bucket in INDEX_BUCKETS:
print(f"INFO Uploading {subdir}/{self.html_name} to {bucket.name}")
bucket.Object(
key=f"{subdir}/{self.html_name}"
).put(
ACL='public-read',
CacheControl='no-cache,no-store,must-revalidate',
ContentType='text/html',
Body=index_html
)
def upload_pep503_htmls(self) -> None:
for subdir in self.subdirs:
index_html = self.to_simple_packages_html(subdir=subdir)
for bucket in INDEX_BUCKETS:
print(f"INFO Uploading {subdir}/index.html to {bucket.name}")
bucket.Object(
key=f"{subdir}/index.html"
).put(
ACL='public-read',
CacheControl='no-cache,no-store,must-revalidate',
ContentType='text/html',
Body=index_html
)
for pkg_name in self.get_package_names(subdir=subdir):
compat_pkg_name = pkg_name.lower().replace("_", "-")
index_html = self.to_simple_package_html(subdir=subdir, package_name=pkg_name)
for bucket in INDEX_BUCKETS:
print(f"INFO Uploading {subdir}/{compat_pkg_name}/index.html to {bucket.name}")
bucket.Object(
key=f"{subdir}/{compat_pkg_name}/index.html"
).put(
ACL='public-read',
CacheControl='no-cache,no-store,must-revalidate',
ContentType='text/html',
Body=index_html
)
def save_libtorch_html(self) -> None:
for subdir in self.subdirs:
print(f"INFO Saving {subdir}/{self.html_name}")
makedirs(subdir, exist_ok=True)
with open(path.join(subdir, self.html_name), mode="w", encoding="utf-8") as f:
f.write(self.to_libtorch_html(subdir=subdir))
def save_pep503_htmls(self) -> None:
for subdir in self.subdirs:
print(f"INFO Saving {subdir}/index.html")
makedirs(subdir, exist_ok=True)
with open(path.join(subdir, "index.html"), mode="w", encoding="utf-8") as f:
f.write(self.to_simple_packages_html(subdir=subdir))
for pkg_name in self.get_package_names(subdir=subdir):
makedirs(path.join(subdir, pkg_name), exist_ok=True)
with open(path.join(subdir, pkg_name, "index.html"), mode="w", encoding="utf-8") as f:
f.write(self.to_simple_package_html(subdir=subdir, package_name=pkg_name))
def compute_sha256(self) -> None:
for obj in self.objects:
if obj.checksum is not None:
continue
print(f"Updating {obj.orig_key} of size {obj.size} with SHA256 checksum")
s3_obj = BUCKET.Object(key=obj.orig_key)
s3_obj.copy_from(CopySource={"Bucket": BUCKET.name, "Key": obj.orig_key},
Metadata=s3_obj.metadata, MetadataDirective="REPLACE",
ACL="public-read",
ChecksumAlgorithm="SHA256")
@classmethod
def has_public_read(cls: Type[S3IndexType], key: str) -> bool:
def is_all_users_group(o) -> bool:
return o.get("Grantee", {}).get("URI") == "http://acs.amazonaws.com/groups/global/AllUsers"
def can_read(o) -> bool:
return o.get("Permission") in ["READ", "FULL_CONTROL"]
acl_grants = CLIENT.get_object_acl(Bucket=BUCKET.name, Key=key)["Grants"]
return any(is_all_users_group(x) and can_read(x) for x in acl_grants)
@classmethod
def grant_public_read(cls: Type[S3IndexType], key: str) -> None:
CLIENT.put_object_acl(Bucket=BUCKET.name, Key=key, ACL="public-read")
@classmethod
def fetch_object_names(cls: Type[S3IndexType], prefix: str) -> List[str]:
obj_names = []
for obj in BUCKET.objects.filter(Prefix=prefix):
is_acceptable = any([path.dirname(obj.key) == prefix] + [
match(
f"{prefix}/{pattern}",
path.dirname(obj.key)
)
for pattern in ACCEPTED_SUBDIR_PATTERNS
]) and obj.key.endswith(ACCEPTED_FILE_EXTENSIONS)
if not is_acceptable:
continue
obj_names.append(obj.key)
return obj_names
def fetch_metadata(self: S3IndexType) -> None:
# Add PEP 503-compatible hashes to URLs to allow clients to avoid spurious downloads, if possible.
with concurrent.futures.ThreadPoolExecutor(max_workers=6) as executor:
for idx, future in {
idx: executor.submit(
lambda key: CLIENT.head_object(
Bucket=BUCKET.name, Key=key, ChecksumMode="Enabled"
),
obj.orig_key,
)
for (idx, obj) in enumerate(self.objects)
if obj.size is None
}.items():
response = future.result()
sha256 = (_b64 := response.get("ChecksumSHA256")) and base64.b64decode(_b64).hex()
# For older files, rely on checksum-sha256 metadata that can be added to the file later
if sha256 is None:
sha256 = response.get("Metadata", {}).get("checksum-sha256")
self.objects[idx].checksum = sha256
if size := response.get("ContentLength"):
self.objects[idx].size = int(size)
def fetch_pep658(self: S3IndexType) -> None:
def _fetch_metadata(key: str) -> str:
try:
response = CLIENT.head_object(
Bucket=BUCKET.name, Key=f"{key}.metadata", ChecksumMode="Enabled"
)
sha256 = base64.b64decode(response.get("ChecksumSHA256")).hex()
return sha256
except botocore.exceptions.ClientError as e:
if e.response["Error"]["Code"] == "404":
return None
raise
with concurrent.futures.ThreadPoolExecutor(max_workers=6) as executor:
metadata_futures = {
idx: executor.submit(
_fetch_metadata,
obj.orig_key,
)
for (idx, obj) in enumerate(self.objects)
}
for idx, future in metadata_futures.items():
response = future.result()
if response is not None:
self.objects[idx].pep658 = response
@classmethod
def from_S3(cls: Type[S3IndexType], prefix: str, with_metadata: bool = True) -> S3IndexType:
prefix = prefix.rstrip("/")
obj_names = cls.fetch_object_names(prefix)
def sanitize_key(key: str) -> str:
return key.replace("+", "%2B")
rc = cls([S3Object(key=sanitize_key(key),
orig_key=key,
checksum=None,
size=None,
pep658=None) for key in obj_names], prefix)
if prefix == "whl/nightly":
rc.objects = rc.nightly_packages_to_show()
if with_metadata:
rc.fetch_metadata()
rc.fetch_pep658()
return rc
@classmethod
def undelete_prefix(cls: Type[S3IndexType], prefix: str) -> None:
paginator = CLIENT.get_paginator("list_object_versions")
for page in paginator.paginate(Bucket=BUCKET.name, Prefix=prefix):
for obj in page.get("DeleteMarkers", []):
if not obj.get("IsLatest"):
continue
obj_key, obj_version_id = obj["Key"], obj["VersionId"]
obj_ver = S3.ObjectVersion(BUCKET.name, obj_key, obj_version_id)
print(f"Undeleting {obj_key} deleted on {obj['LastModified']}")
obj_ver.delete()
def create_parser() -> argparse.ArgumentParser:
parser = argparse.ArgumentParser("Manage S3 HTML indices for PyTorch")
parser.add_argument(
"prefix",
type=str,
choices=PREFIXES + ["all"]
)
parser.add_argument("--do-not-upload", action="store_true")
parser.add_argument("--compute-sha256", action="store_true")
return parser
def main() -> None:
parser = create_parser()
args = parser.parse_args()
action = "Saving indices" if args.do_not_upload else "Uploading indices"
if args.compute_sha256:
action = "Computing checksums"
prefixes = PREFIXES if args.prefix == 'all' else [args.prefix]
for prefix in prefixes:
generate_pep503 = prefix.startswith("whl")
print(f"INFO: {action} for '{prefix}'")
stime = time.time()
idx = S3Index.from_S3(prefix=prefix, with_metadata=generate_pep503 or args.compute_sha256)
etime = time.time()
print(f"DEBUG: Fetched {len(idx.objects)} objects for '{prefix}' in {etime-stime:.2f} seconds")
if args.compute_sha256:
idx.compute_sha256()
elif args.do_not_upload:
if generate_pep503:
idx.save_pep503_htmls()
else:
idx.save_libtorch_html()
else:
if generate_pep503:
idx.upload_pep503_htmls()
else:
idx.upload_libtorch_html()
if __name__ == "__main__":
main()