PRIME-RL is a framework for large-scale asynchronous reinforcement learning. It is designed to be easy-to-use and hackable, yet capable of scaling to 1000+ GPUs. Beyond that, here is why we think you might like it:
- Integrates natively with
verifiersenvironments via the Environments Hub - Supports end-to-end post-training, including SFT and RL training and evals
- Multi-node deployment with FSDP2 training and vLLM inference backend
- Designed for asynchronous training in decentralized settings
- Hackable, modular and extensible by nature
We develop and test on NVIDIA RTX 3090/4090/5090, A100, H100, H200, and B200. If your setup fails, please create an issue.
Currently, you need at least one NVIDIA GPU to use PRIME-RL. If you don't already have access to one, we recommend our compute platform for everything from renting on-demand single GPUs for developing, debugging and small ablations, to reserving 1000+ GPU clusters for production-scale training.
Set up PRIME-RL in a single command.
curl -sSL https://raw.githubusercontent.com/PrimeIntellect-ai/prime-rl/main/scripts/install.sh | bashManual Setup
- Clone the repository
git clone https://github.com/PrimeIntellect-ai/prime-rl.git
cd prime-rl- Install uv
curl -LsSf https://astral.sh/uv/install.sh | sh
source $HOME/.local/bin/env- Install dependencies from the lock file
uv syncValidate your environment setup
- Check that the environment uses Python 3.12
uv run python -V- Check that
flash-attnis installed
uv run python -c "import flash_attn"- Check that you can run SFT trainer (this requires 1 GPU)
uv run sft @ configs/debug/sft/train.toml- Check that you can run the RL trainer (this requires 1 GPU)
uv run trainer @ configs/debug/rl/train.toml- Check that you can run the inference server (this requires 1 GPU)
uv run inference @ configs/debug/infer.tomlKeep the inference server running in the background for the next steps.
5.1. Check that you can run the orchestrator against the inference server
uv run orchestrator @ configs/debug/orch.toml5.2. Check that you can run evals against the inference server
uv run eval @ configs/debug/eval.toml- If you want to log your runs to W&B, log in
uv run wandb login
# Or set `export WANDB_API_KEY=...`- If you require gated/ private models or datasets from HuggingFace, log in
uv run hf auth login
# Or set `export HF_TOKEN=...`We provide end-to-end training examples in the examples directory to highlight features of the framework and guide you through the process of training your own models.
- Reverse Text: Train
Qwen3-0.6Bto reverse a small chunk of text. Demonstrates tiny-scale single-turn SFT and RL training. Can be trained on a single consumer GPU in a few minutes, and is ideal for getting started. - Wordle: Train
Qwen3-1.7Bto play Wordle. A fun example of multi-turn SFT and RL training. Can be trained on a 2-4 H100 GPUs in a few hours. Ideal for exploring the multi-turn training capabilities of the framework. - Alphabet Sort: Train
Qwen3-4B-Instruct-2507to sort names alphabetically. Demonstrates multi-turn RL training via LoRA without SFT warmup. Can be trained on a single H100 GPU in just over an hour. Ideal for exploring LoRA-based training. - More to come...
Check out the docs directory for in-depth guides on how to use PRIME-RL.
- Entrypoints - Overview of the main components (orchestrator, trainer, inference) and how to run SFT, RL, and evals
- Configs - Configuration system using TOML files, CLI arguments, and environment variables
- Environments - Installing and using verifiers environments from the Environments Hub
- Async Training - Understanding asynchronous off-policy training and step semantics
- Logging - Logging with loguru, torchrun, and Weights & Biases
- Checkpointing - Saving and resuming training from checkpoints
- Benchmarking - Performance benchmarking and throughput measurement
- Deployment - Training deployment on single-GPU, multi-GPU, and multi-node clusters
- Troubleshooting - Common issues and their solutions
We warmly welcome community contributions! We use issues to track bugs, feature requests, and share our internal roadmap. If you encounter bugs, have pain points during development, or have ideas for new features, please open an issue.
Contributions are welcome via PR. Please follow these guidelines:
- Install the pre-commit hooks to ensure your code is formatted correctly.
- Please keep your PR in "Draft" until it is ready for review.
- If your PR resolves an issue, please link the issue in the PR description
- If you can, try running the test suite locally to ensure your changes are working as expected.
Please install the pre-commit hooks to ensure your code is formatted correctly.
uv run pre-commit installRun the full test suite
uv run pytest -vTo run unit tests, run
uv run pytest tests/unit -vTo run integration tests, run
uv run pytest tests/integration -vTo run CPU-only tests, use the inverse of the gpu marker:
uv run pytest -v -m "not gpu"This project is licensed under the Apache 2.0 license, as found in the License file.
If you find our work useful, feel free to cite it using
@misc{primeintellect2025prime-rl,
author = {Prime Intellect},
title = {PRIME-RL},
url = {https://github.com/PrimeIntellect-ai/prime-rl},
year = {2025}
}