pretraining
Here are 222 public repositories matching this topic...
Official repository of OFA (ICML 2022). Paper: OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning Framework
-
Updated
Apr 24, 2024 - Python
mPLUG-Owl: The Powerful Multi-modal Large Language Model Family
-
Updated
Apr 2, 2025 - Python
Papers about pretraining and self-supervised learning on Graph Neural Networks (GNN).
-
Updated
Feb 2, 2024 - Python
[ICLR'23 Spotlight🔥] The first successful BERT/MAE-style pretraining on any convolutional network; Pytorch impl. of "Designing BERT for Convolutional Networks: Sparse and Hierarchical Masked Modeling"
-
Updated
Jan 23, 2024 - Python
Recent Advances in Vision and Language PreTrained Models (VL-PTMs)
-
Updated
Aug 19, 2022
EntitySeg Toolbox: Towards Open-World and High-Quality Image Segmentation
-
Updated
Nov 30, 2023 - Jupyter Notebook
X-modaler is a versatile and high-performance codebase for cross-modal analytics(e.g., image captioning, video captioning, vision-language pre-training, visual question answering, visual commonsense reasoning, and cross-modal retrieval).
-
Updated
Feb 27, 2023 - Python
Official Repository for the Uni-Mol Series Methods
-
Updated
May 29, 2025 - Python
Pretraining and inference code for a large-scale depth-recurrent language model
-
Updated
Sep 5, 2025 - Python
【ICLR 2024🔥】 Extending Video-Language Pretraining to N-modality by Language-based Semantic Alignment
-
Updated
Mar 25, 2024 - Python
Official Pytorch Implementation of: "ImageNet-21K Pretraining for the Masses"(NeurIPS, 2021) paper
-
Updated
Jan 11, 2023 - Python
A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites
-
Updated
Jul 19, 2025
The official implementation of MARS: Unleashing the Power of Variance Reduction for Training Large Models
-
Updated
Oct 6, 2025 - Python
Best practice for training LLaMA models in Megatron-LM
-
Updated
Jan 2, 2024 - Python
Official repository for "Craw4LLM: Efficient Web Crawling for LLM Pretraining"
-
Updated
Feb 24, 2025 - Python
Saprot: Protein Language Model with Structural Alphabet (AA+3Di)
-
Updated
Oct 9, 2025 - Python
PITI: Pretraining is All You Need for Image-to-Image Translation
-
Updated
Jun 2, 2024 - Python
Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Train your own 8B/14B LLaVA-training-like MLLM on RTX3090/4090 24GB.
-
Updated
Mar 10, 2025 - Jupyter Notebook
Improve this page
Add a description, image, and links to the pretraining topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the pretraining topic, visit your repo's landing page and select "manage topics."