Skip to content

Commit 77fa0bc

Browse files
Merge branch 'volcengine:main' into main
2 parents 5415881 + 0615352 commit 77fa0bc

File tree

85 files changed

+4244
-410
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

85 files changed

+4244
-410
lines changed

.github/workflows/e2e_genrm_remote.yml

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -66,6 +66,7 @@ on:
6666
- ".github/workflows/e2e_genrm_remote.yml"
6767
- "examples/data_preprocess/gsm8k.py"
6868
- "tests/special_e2e/run_genrm_remote.sh"
69+
- "tests/special_e2e/generation/run_gen_qwen05_server.sh"
6970

7071
# Cancel jobs on the same ref if a new one is triggered
7172
concurrency:
@@ -120,6 +121,8 @@ jobs:
120121
run: |
121122
ray stop --force
122123
bash tests/special_e2e/run_genrm_remote.sh
124+
ray stop --force
125+
bash tests/special_e2e/generation/run_gen_qwen05_server.sh
123126
124127
cleanup:
125128
runs-on: ubuntu-latest

.github/workflows/e2e_sft.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -91,7 +91,7 @@ jobs:
9191
e2e_sft:
9292
needs: setup
9393
runs-on: ["${{ needs.setup.outputs.runner-label || 'L20x8' }}"]
94-
timeout-minutes: 25 # Increase this timeout value as needed
94+
timeout-minutes: 30 # Increase this timeout value as needed
9595
env:
9696
HTTP_PROXY: ${{ secrets.PROXY_HTTP }}
9797
HTTPS_PROXY: ${{ secrets.PROXY_HTTPS }}

.github/workflows/model.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -208,6 +208,7 @@ jobs:
208208
209209
- name: Running mcore engine tests on 8 L20 GPUs
210210
run: |
211+
ray stop --force
211212
pytest -s -x tests/models/test_engine.py
212213
213214
cleanup:

README.md

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -238,6 +238,9 @@ verl is inspired by the design of Nemo-Aligner, Deepspeed-chat and OpenRLHF. The
238238
- [Vision-SR1](https://github.com/zli12321/Vision-SR1): Self-Rewarding Vision-Language Model via Reasoning Decomposition ![GitHub Repo stars](https://img.shields.io/github/stars/zli12321/Vision-SR1)
239239
- [SimpleVLA-RL](https://github.com/PRIME-RL/SimpleVLA-RL): SimpleVLA-RL: A Simple yet Effective Vision-Language Action Model for Reinforcement Learning ![GitHub Repo stars](https://img.shields.io/github/stars/PRIME-RL/SimpleVLA-RL)
240240
- [Table-R1](https://github.com/Table-R1/Table-R1): Table-R1: Inference-Time Scaling for Table Reasoning ![GitHub Repo stars](https://img.shields.io/github/stars/Table-R1/Table-R1)
241+
- [Revisual-R1](https://github.com/CSfufu/Revisual-R1): Revisual-R1: Advancing Multimodal Reasoning From Optimized Cold Start to Staged Reinforcement Learning ![GitHub Repo stars](https://img.shields.io/github/stars/CSfufu/Revisual-R1)
242+
- [ARES](https://github.com/shawn0728/ARES): ARES: Multimodal Adaptive Reasoning via Difficulty-Aware Token-Level Entropy Shaping ![GitHub Repo stars](https://img.shields.io/github/stars/shawn0728/ARES)
243+
- [Meta-Bandit-LLM](https://github.com/sanxing-chen/meta-bandit-llm): Meta-Bandit-LLM: Long-horizon multiturn interactive training for meta-bandit agents ![GitHub Repo stars](https://img.shields.io/github/stars/sanxing-chen/meta-bandit-llm)
241244

242245
and many more awesome work listed in [recipe](recipe/README.md).
243246

docker/README.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -36,6 +36,8 @@ For vLLM with FSDP, please refer to [hiyouga/verl](https://hub.docker.com/r/hiyo
3636

3737
For SGLang with FSDP, please refer to [ocss884/verl-sglang](https://hub.docker.com/r/ocss884/verl-sglang) repository and the latest version is ``ocss884/verl-sglang:ngc-th2.6.0-cu126-sglang0.4.6.post5`` which is provided by SGLang RL Group.
3838

39+
For latest vLLM with Megatron, please refer to [iseekyan/verl](https://hub.docker.com/r/iseekyan/verl) repository and the latest version is ``iseekyan/verl:nemo.gptoss_vllm0.11.0``.
40+
3941
See files under ``docker/`` for NGC-based image or if you want to build your own.
4042

4143
Note that For aws instances with EFA net interface (Sagemaker AI Pod), you need to install EFA driver as shown in ``docker/Dockerfile.extenstion.awsefa``
Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,15 @@
1+
FROM nvcr.io/nvidia/nemo:25.07.gpt_oss
2+
3+
RUN git clone -b v0.11.0 --depth 1 https://github.com/vllm-project/vllm.git /opt/vllm
4+
5+
RUN pip install setuptools_scm
6+
7+
RUN cd /opt/vllm && pip install --no-deps --no-build-isolation --no-cache-dir -e .
8+
9+
RUN pip install cbor2 setproctitle blake3 openai_harmony pybase64 msgspec partial_json_parser py-cpuinfo diskcache gguf
10+
11+
RUN pip install --upgrade transformers tokenizers
12+
13+
RUN pip install codetiming tensordict mathruler pylatexenc
14+
15+
RUN pip3 install --no-cache-dir mbridge

0 commit comments

Comments
 (0)