Skip to content

Commit 3cee68a

Browse files
authored
Merge pull request #1297 from ELC/pytorchconf-2024
Add PytorchConf 2024
2 parents 5d916f4 + af7be93 commit 3cee68a

File tree

87 files changed

+2178
-0
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

87 files changed

+2178
-0
lines changed

pytorchconf-2024/category.json

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,3 @@
1+
{
2+
"title": "PyTorch Conference 2024"
3+
}
Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
{
2+
"description": "A Distributed Stateful Dataloader for Large-Scale Pretraining - Davis Wertheimer, IBM & Linsong Chu, IBM Research\n\nLarge-scale model pretraining crucially relies on specialized and dedicated dataloaders that can, for example, partition and stream data asynchronously across multiple processes and physical nodes. In this talk we discuss one of the torch-native dataloaders we built and use at IBM Research for addressing these needs. Intended for use in large-scale model pretraining, particularly in research settings where rapid iteration between datasets may be required, our dataloader is distributed, stateful, checkpointable, composable and rescalable \u2013 while remaining a simple extension of the existing PyTorch dataloading framework. It automatically and invisibly handles data sharding, shuffling, subdataset weighting, checkpoint saving and loading, and custom user-defined preprocessing functions, with minimal overhead and high throughput. We discuss these properties and how we achieved them, such as reducing overhead by implementing a custom LCG random number generator, and demonstrate proof of concept on production-scale training of a 7B parameter Llama model over 4 trillion tokens.",
3+
"duration": 1433,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Davis Wertheimer",
14+
"Linsong Chu"
15+
],
16+
"tags": [],
17+
"thumbnail_url": "https://i.ytimg.com/vi_webp/VtT4rdph4Qs/maxresdefault.webp",
18+
"title": "A Distributed Stateful Dataloader for Large-Scale Pretraining",
19+
"videos": [
20+
{
21+
"type": "youtube",
22+
"url": "https://www.youtube.com/watch?v=VtT4rdph4Qs"
23+
}
24+
]
25+
}
Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
{
2+
"description": "Blobs to Clips: Efficient End-to-End Video Data Loading - Andrew Ho & Ahmad Sharif, Meta\n\nThe PyTorch team has improved training speed by an order of magnitude for teams at Meta working on Small-to-Large-Scale MultiModal Video models. In this talk we\u2019ll share our learnings on reducing GPU starvation by overcoming data loading challenges such as dealing with large distributed datasets, worker imbalance, compute-bottlenecks due to parallel video decoding and sampling, checkpointing, and debuggability. As part of our commitment to open-source, we are releasing a new decoding library and updating existing PyTorch libraries on GitHub, and invite feedback and contributions from the community.",
3+
"duration": 1544,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Andrew Ho",
14+
"Ahmad Sharif"
15+
],
16+
"tags": [],
17+
"thumbnail_url": "https://i.ytimg.com/vi_webp/-cBplHNM4RA/maxresdefault.webp",
18+
"title": "Blobs to Clips: Efficient End-to-End Video Data Loading",
19+
"videos": [
20+
{
21+
"type": "youtube",
22+
"url": "https://www.youtube.com/watch?v=-cBplHNM4RA"
23+
}
24+
]
25+
}
Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
{
2+
"description": "Building PyTorch Computer Vision Algorithms for 100 Skin Shades - Emmanuel Acheampong, roboMUA\n\nAt roboMUA we're leading the charge in building predictive AI models for diverse skin shades with the use of Convolutional Neural Networks (CNNs), and harnessing the power of Generative Adversarial Networks (GANs) specifically for generating realistic images of black hairstyles. Our session showcases PyTorch's versatility in both predictive and generative tasks, offering a comprehensive approach to inclusive AI. For predictive AI models, we leverage PyTorch's flexible framework to develop CNNs. Through innovative techniques in feature engineering and model architecture design, we demonstrate how PyTorch enables accurate prediction across 100 skin shades. Simultaneously, we showcase the transformative potential of GANs in the realm of black hairstyles. By training GANs on a curated dataset of diverse hair textures and styles, we illustrate how PyTorch facilitates the generation of lifelike images that celebrate the beauty and diversity of black hair. Attendees will gain insights into the data preprocessing, model training, and evaluation processes and and learn how PyTorch empowers developers to build inclusive solutions.",
3+
"duration": 882,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Emmanuel Acheampong"
14+
],
15+
"tags": [],
16+
"thumbnail_url": "https://i.ytimg.com/vi_webp/eGZ3-8ZbFj0/maxresdefault.webp",
17+
"title": "Building PyTorch Computer Vision Algorithms for 100 Skin Shades",
18+
"videos": [
19+
{
20+
"type": "youtube",
21+
"url": "https://www.youtube.com/watch?v=eGZ3-8ZbFj0"
22+
}
23+
]
24+
}
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
{
2+
"description": "Building Scientific Computing Infrastructure Software with the PyTorch Ecosystem - Bharath Ramsundar, Deep Forest Sciences\n\nThe DeepChem library is a scientific computing library that implements deep learning infrastructure for drug discovery, materials discovery, and biology. The DeepChem community is one of the largest scientific open source projects built in PyTorch, with over 5K stars on Github and thousands of citations. The DeepChem community has learned a number of useful lessons for building and maintaining high quality scientific code built on top of PyTorch. In this talk, I will share our learnings with the PyTorch community and also highlight opportunities for improving scientific support in the ecosystem.",
3+
"duration": 1685,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Bharath Ramsundar"
14+
],
15+
"tags": [],
16+
"thumbnail_url": "https://i.ytimg.com/vi_webp/Ru-CDK_79NQ/maxresdefault.webp",
17+
"title": "Building Scientific Computing Infrastructure Software with the PyTorch Ecosystem",
18+
"videos": [
19+
{
20+
"type": "youtube",
21+
"url": "https://www.youtube.com/watch?v=Ru-CDK_79NQ"
22+
}
23+
]
24+
}
Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
{
2+
"description": "Data-Dependent Shapes in PT2 - Edward Yang, Meta\n\nData-dependent shapes are ubiquitous whenever you want to take advantage of sparsity in your data representation, whether it is in recommendation systems, mixture of experts or other use cases. We have made a lot of improvements to torch.compile's support for capturing and compiling data dependent shapes, but they also require some user knowledge to work with effectively. This talk will give an overview of PT2's facilities for data dependent compute and how to use them effectively.",
3+
"duration": 1524,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Edward Yang"
14+
],
15+
"tags": [],
16+
"thumbnail_url": "https://i.ytimg.com/vi_webp/VYuFJUfyWSU/maxresdefault.webp",
17+
"title": "Data-Dependent Shapes in PT2",
18+
"videos": [
19+
{
20+
"type": "youtube",
21+
"url": "https://www.youtube.com/watch?v=VYuFJUfyWSU"
22+
}
23+
]
24+
}
Lines changed: 29 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,29 @@
1+
{
2+
"description": "DL Compiler Panel Discussion - Philip Tillet, OpenAI; Jason Ansel, Meta; Jacques Pienaar, Google; Tianqi Chen, CMU & OctoAI; Mikhail Zolotukhin, Modular; Peng Wu, Meta\n\nSince the release of PyTorch 2 in 2023, torch.compile() has spurred significant new thinking around DL compiler designs at the framework level. In this session, we invite leaders in this space to share their insights based on real experiences of building DL compilers \u2013 Triton, TorchInductor, Halide, TVM, OpenXLA, and Mojo \u2013 and growing their ecosystems. We also invite a \u2018compiler user representative,\u2019 together.ai, to share their recent journey of redesigning the LLM inference stack around torch.compile(). Each leader will give a 10-minute lightning talk and an engaging panel discussion.",
3+
"duration": 2131,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Philip Tillet",
14+
"Jason Ansel",
15+
"Jacques Pienaar",
16+
"Tianqi Chen",
17+
"Mikhail Zolotukhin",
18+
"Peng Wu"
19+
],
20+
"tags": [],
21+
"thumbnail_url": "https://i.ytimg.com/vi_webp/YWDzHGx8PrY/maxresdefault.webp",
22+
"title": "DL Compiler Panel Discussion",
23+
"videos": [
24+
{
25+
"type": "youtube",
26+
"url": "https://www.youtube.com/watch?v=YWDzHGx8PrY"
27+
}
28+
]
29+
}
Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
{
2+
"description": "ExecuTorch Beta and on-Device Generative AI Support - Mergen Nachin & Mengtao (Martin) Yuan, Meta\n\nDuring this session, we will discuss real-life case studies focusing on the productionization of PyTorch models onto edge devices and welcome the community to begin adopting ExecuTorch. Since announcing the ExecuTorch MVP at the previous PTC, we have made significant progress in terms of stability, model coverage, accelerator performance, and developer experience, reaching a milestone that marks the transition to beta status. In addition to the above improvements, we continue to support generative AI models. Since the alpha launch that initially enabled support for LLama2/3 models, we have now expanded our capabilities to include multimodal use cases and developed mobile demo apps showcasing these new features.",
3+
"duration": 1213,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Mergen Nachin",
14+
"Mengtao Yuan"
15+
],
16+
"tags": [],
17+
"thumbnail_url": "https://i.ytimg.com/vi_webp/45yNTi7c1Q0/maxresdefault.webp",
18+
"title": "ExecuTorch Beta and on-Device Generative AI Support",
19+
"videos": [
20+
{
21+
"type": "youtube",
22+
"url": "https://www.youtube.com/watch?v=45yNTi7c1Q0"
23+
}
24+
]
25+
}
Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
{
2+
"description": "Hacks to Make LLM Training Faster - Daniel Han, Unsloth AI\n\nAs open-source LLMs have become more capable, a substantial ecosystem has developed around the fine-tuning of these models. A thriving community of researchers, developers, practitioners and hobbyists has emerged which focuses on topics ranging from memory efficiency, parameter-efficient fine-tuning and quantization to performance at scale and reproducible evaluations. The goal of this mini-summit is to bring this community together to discuss ideas, share knowledge and build connections.\n\nThe agenda features a keynote from Joe Spisak on the state of the Llama ecosystem followed by invited talks from the founders of Axolotl, Unsloth and torchtune. We conclude the summit with a riveting discussion on what\u2019s next for LLMs, fine-tuning and the PyTorch ecosystem with a fabulous panel of experts - Tim Dettmers (author of bitsandbytes and QLoRA), Hailey Schoelkopf (maintainer of LM Eval Harness at EleutherAI), Aakanksha Chowdhery (Lead author on PaLM and Gemini) and Alexis Conneau (Research Lead at OpenAI)",
3+
"duration": 477,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Daniel Han"
14+
],
15+
"tags": [],
16+
"thumbnail_url": "https://i.ytimg.com/vi_webp/PdtKkc5jB4g/maxresdefault.webp",
17+
"title": "Hacks to Make LLM Training Faster",
18+
"videos": [
19+
{
20+
"type": "youtube",
21+
"url": "https://www.youtube.com/watch?v=PdtKkc5jB4g"
22+
}
23+
]
24+
}
Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
{
2+
"description": "[HALIDE] A Halide Backend for TorchInductor - Jason Ansel, Meta\n\nThis talk will focus on a new Halide backend for TorchInductor, which is in addition to the existing Triton and C++ backends. The Halide backend is meant to serve as a reference backend to make it easier to extend TorchInductor to support new backend compilers and hardware devices. Halide has been the inspiration (either in ideas or through forking) of numerous other compiler projects, so it is a good starting point for adding new backends that follow a Halide-like model.",
3+
"duration": 489,
4+
"language": "eng",
5+
"recorded": "2024-09-18",
6+
"related_urls": [
7+
{
8+
"label": "Conference Website",
9+
"url": "https://pytorch.org/event/pytorch-conference-2024/"
10+
}
11+
],
12+
"speakers": [
13+
"Jason Ansel"
14+
],
15+
"tags": [],
16+
"thumbnail_url": "https://i.ytimg.com/vi_webp/hCvlWZEXRrg/maxresdefault.webp",
17+
"title": "[HALIDE] A Halide Backend for TorchInductor",
18+
"videos": [
19+
{
20+
"type": "youtube",
21+
"url": "https://www.youtube.com/watch?v=hCvlWZEXRrg"
22+
}
23+
]
24+
}

0 commit comments

Comments
 (0)