Change the repository type filter
All
Repositories list
32 repositories
ColossalAI
PublicMaking large AI models cheaper, faster and more accessibleTensorRT-LLM
PublicTensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and support state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT LLM also contains components to create Python and C++ runtimes that orchestrate the inference execution in performant way.TensorRT-Model-Optimizer
PublicA unified library of state-of-the-art model optimization techniques like quantization, pruning, distillation, speculative decoding, etc. It compresses deep learning models for downstream deployment frameworks like TensorRT-LLM or TensorRT to optimize inference speed.Open-Sora
PublicOpen-Sora-Demo
PublicColossalAI-Platform-CLI
Public- A Python library transfers PyTorch tensors between CPU and NVMe
transformers
PublicFastFold
PublicOptimizing AlphaFold Training and Inference on GPU ClustersCANN-Installer
PublicSwiftInfer
PublicEfficient AI Inference & ServingElixir
PublicElixir: Train a Large Language Model on a Small GPU ClusterColossalAI-Examples
Public archivemmdetection-examples
Public archiveGPT-Demo
Public archiveTitans
Public archiveOPT-Benchmark
Public archiveCachedEmbedding
PublicA memory efficient DLRM training solution using ColossalAISkyComputing
Public archive