Skip to content

Commit 77814f1

Browse files
ISEEKYANwuxibin89
andauthored
[env] feat: update docker file building schema, from VLLM base images (volcengine#3937)
### What does this PR do? Use VLLM/SGLANG images as base to build stable images. Would like to build nightly env image with all the latest version of main components including CUDA/CUDNN/VLLM/SGLANG/TransformerEngine/Megatron. ### Test `docker/Dockerfile.stable.vllm011` is tested with `examples/grpo_trainer/run_qwen3_vl-30b-megatron.sh`. sglang image is not tested yet. > [!IMPORTANT] > Please check all the following items before requesting a review, otherwise the reviewer might deprioritize this PR for review. - [ ] Read the [Contribute Guide](https://github.com/volcengine/verl/blob/main/CONTRIBUTING.md). - [ ] Apply [pre-commit checks](https://github.com/volcengine/verl/blob/main/CONTRIBUTING.md#code-linting-and-formatting): `pre-commit install && pre-commit run --all-files --show-diff-on-failure --color=always` - [ ] Add / Update [the documentation](https://github.com/volcengine/verl/tree/main/docs). - [ ] Add unit or end-to-end test(s) to [the CI workflow](https://github.com/volcengine/verl/tree/main/.github/workflows) to cover all the code. If not feasible, explain why: ... - [ ] Once your PR is ready for CI, send a message in [the `ci-request` channel](https://verl-project.slack.com/archives/C091TCESWB1) in [the `verl` Slack workspace](https://join.slack.com/t/verl-project/shared_invite/zt-3855yhg8g-CTkqXu~hKojPCmo7k_yXTQ). (If not accessible, please try [the Feishu group (飞书群)](https://applink.larkoffice.com/client/chat/chatter/add_by_link?link_token=772jd4f1-cd91-441e-a820-498c6614126a).) --------- Co-authored-by: wuxibin <[email protected]>
1 parent 703a078 commit 77814f1

26 files changed

+128
-57
lines changed

.github/workflows/.deprecate/e2e_eval_aime24.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ permissions:
8888
contents: read
8989

9090
env:
91-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
91+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
9292
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
9393

9494
jobs:

.github/workflows/cpu_unit_tests.yml

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -68,16 +68,17 @@ jobs:
6868
NO_PROXY: "localhost,127.0.0.1,hf-mirror.com"
6969
HF_ENDPOINT: "https://hf-mirror.com"
7070
HF_HUB_ENABLE_HF_TRANSFER: "0" # This is more stable
71+
TORCH_COMPILE_DISABLE: 1
72+
TORCHINDUCTOR_DISABLE: 1
7173
container:
72-
image: verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2
74+
image: verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7
7375
steps:
7476
- uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
7577
with:
7678
fetch-depth: 0
7779
- name: Install the current repository
7880
run: |
79-
pip install -e .[test,prime,geo]
80-
pip install --upgrade "ray>=2.40.0" pillow
81+
pip install -e .[test,geo]
8182
- name: Download datasets
8283
run: |
8384
huggingface-cli download verl-team/gsm8k-v0.4.1 --repo-type dataset --local-dir ~/verl-data/gsm8k

.github/workflows/e2e_dapo.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ permissions:
8484
contents: read
8585

8686
env:
87-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
87+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
8888
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
8989

9090
jobs:

.github/workflows/e2e_fully_async_policy.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ permissions:
8484
contents: read
8585

8686
env:
87-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
87+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
8888
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
8989
TRANSFORMERS_VERSION: "4.56.2"
9090

.github/workflows/e2e_genrm_remote.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,7 @@ permissions:
7878
contents: read
7979

8080
env:
81-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
81+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
8282
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
8383

8484
jobs:

.github/workflows/e2e_one_step_off_policy.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ permissions:
8484
contents: read
8585

8686
env:
87-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
87+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
8888
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
8989
TRANSFORMERS_VERSION: "4.56.2"
9090

.github/workflows/e2e_ppo_trainer_megatron_sglang_2.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -227,7 +227,7 @@ jobs:
227227
ray stop --force
228228
TRAIN_FILES=$HOME/data/geo3k/train.parquet VAL_FILES=$HOME/data/geo3k/test.parquet \
229229
MAX_PROMPT_LEN=1536 MAX_RESPONSE_LEN=1536 \
230-
MODEL_ID=Qwen/Qwen2-VL-2B-Instruct \
230+
MODEL_ID=Qwen/Qwen2.5-VL-3B-Instruct \
231231
ADV_ESTIMATOR=grpo RM_PAD=True USE_KL=True ENABLE_CHUNKED_PREFILL=False \
232232
ENGINE=sglang GPU_MEMORY_UTILIZATION=0.6 ACTOR_FSDP_PARAM_OFFLOAD=True \
233233
ACTOR_FSDP_OPTIMIZER_OFFLOAD=True REF_FSDP_PARAM_OFFLOAD=True \

.github/workflows/e2e_ppo_trainer_megatron_vllm.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,7 @@ permissions:
8585
contents: read
8686

8787
env:
88-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
88+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
8989
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
9090
TRANSFORMERS_VERSION: "4.56.2"
9191

.github/workflows/e2e_ppo_trainer_megatron_vllm_2.yml

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,7 @@ permissions:
8585
contents: read
8686

8787
env:
88-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
88+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
8989
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
9090
TRANSFORMERS_VERSION: "4.56.2"
9191

@@ -164,8 +164,8 @@ jobs:
164164
run: |
165165
ray stop --force
166166
ADV_ESTIMATOR=grpo USE_DUMMY_MODEL=True DUMMY_MODEL_CONFIG_PATH=tests/special_e2e/ppo_trainer/expert_parallel/qwen2moe_minimal.json \
167-
PPO_MAX_TOKEN_LEN=512 FWD_MAX_TOKEN_LEN=512 \
168-
MAX_PROMPT_LENGTH=256 MAX_RESPONSE_LENGTH=256 \
167+
PPO_MAX_TOKEN_LEN=1024 FWD_MAX_TOKEN_LEN=1024 \
168+
MAX_PROMPT_LENGTH=512 MAX_RESPONSE_LENGTH=512 \
169169
MODEL_ID=Qwen/Qwen1.5-MoE-A2.7B-Chat USE_MBRIDGE=True \
170170
COMMON_PP=2 COMMON_VPP=null COMMON_CP=1 COMMON_TP=4 COMMON_EP=4 COMMON_ETP=1 INFER_TP=8 \
171171
USE_DIST_CKPT=True ALL_OFFLOAD=True SKIP_SAVE_HF_MODEL=1 bash tests/special_e2e/run_ppo_trainer_megatron.sh
@@ -374,7 +374,7 @@ jobs:
374374
ray stop --force
375375
TRAIN_FILES=$HOME/data/geo3k/train.parquet VAL_FILES=$HOME/data/geo3k/test.parquet \
376376
MAX_PROMPT_LEN=1536 MAX_RESPONSE_LEN=1536 \
377-
MODEL_ID=Qwen/Qwen2-VL-2B-Instruct \
377+
MODEL_ID=Qwen/Qwen2.5-VL-3B-Instruct \
378378
ADV_ESTIMATOR=grpo RM_PAD=True USE_KL=True ENABLE_CHUNKED_PREFILL=False \
379379
SP_SIZE=2 \
380380
bash tests/special_e2e/ppo_trainer/run_function_reward.sh
@@ -384,7 +384,7 @@ jobs:
384384
ray stop --force
385385
TRAIN_FILES=$HOME/data/geo3k/train.parquet VAL_FILES=$HOME/data/geo3k/test.parquet \
386386
MAX_PROMPT_LEN=1536 MAX_RESPONSE_LEN=1536 \
387-
MODEL_ID=Qwen/Qwen2-VL-2B-Instruct \
387+
MODEL_ID=Qwen/Qwen2.5-VL-3B-Instruct \
388388
ADV_ESTIMATOR=gae RM_PAD=True USE_KL=True ENABLE_CHUNKED_PREFILL=False \
389389
SP_SIZE=2 \
390390
bash tests/special_e2e/ppo_trainer/run_function_reward.sh
@@ -393,7 +393,7 @@ jobs:
393393
ray stop --force
394394
TRAIN_FILES=$HOME/data/geo3k/train.parquet VAL_FILES=$HOME/data/geo3k/test.parquet \
395395
MAX_PROMPT_LEN=1536 MAX_RESPONSE_LEN=1536 \
396-
MODEL_ID=Qwen/Qwen2-VL-2B-Instruct \
396+
MODEL_ID=Qwen/Qwen2.5-VL-3B-Instruct \
397397
ADV_ESTIMATOR=grpo RM_PAD=True USE_KL=True ENABLE_CHUNKED_PREFILL=False \
398398
SP_SIZE=2 \
399399
LORA_RANK=32 LORA_EXCLUDE=".*visual.*" \

.github/workflows/model.yml

Lines changed: 5 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ concurrency:
6464

6565

6666
env:
67-
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:app-verl0.5-transformers4.55.4-vllm0.10.0-mcore0.13.0-te2.2"
67+
IMAGE: "verl-ci-cn-beijing.cr.volces.com/verlai/verl:vllm011.dev7"
6868
DYNAMIC_RUNNER_ENDPOINT: "https://sd10g3clalm04ug7alq90.apigateway-cn-beijing.volceapi.com/runner"
6969

7070
jobs:
@@ -99,7 +99,7 @@ jobs:
9999
fetch-depth: 0
100100
- name: Install the current repository and upgrade to latest transformers(4.54.0)/flash_attn, transformers 4.55.0 has strange behavior with model backward
101101
run: |
102-
pip3 install --no-deps -e .[test]
102+
pip3 install -e .[test]
103103
pip3 install --upgrade transformers
104104
- name: Running rmpad model tests on 8 L20 GPUs + flash_attn 2.5.8
105105
run: |
@@ -147,8 +147,7 @@ jobs:
147147
fetch-depth: 0
148148
- name: Install the current repository and upgrade to latest transformers/flash_attn
149149
run: |
150-
pip3 install --no-deps -e .[test]
151-
pip3 install --upgrade transformers
150+
pip3 install -e .[test]
152151
- name: Running FSDP2 rmpad model tests on 8 L20 GPUs + latest flash_attn
153152
run: |
154153
STRATEGY=fsdp2 torchrun --nproc_per_node=8 tests/special_distributed/test_fsdp_ckpt.py
@@ -169,8 +168,7 @@ jobs:
169168
fetch-depth: 0
170169
- name: Install the current repository
171170
run: |
172-
pip3 install --no-deps -e .[test]
173-
pip install --upgrade "huggingface_hub[cli]"
171+
pip3 install -e .[test]
174172
# - name: Download model config files
175173
# run: |
176174
# hf download Qwen/Qwen2.5-7B config.json --local-dir $HOME/configs/Qwen/Qwen2.5-7B
@@ -199,9 +197,7 @@ jobs:
199197
fetch-depth: 0
200198
- name: Install the current repository
201199
run: |
202-
pip3 install --no-deps -e .[test]
203-
pip3 install --upgrade tensordict transformers
204-
pip install --upgrade "huggingface_hub[cli]"
200+
pip3 install -e .[test]
205201
- name: Download model config files
206202
run: |
207203
hf download Qwen/Qwen2.5-0.5B-Instruct --local-dir $HOME/models/Qwen/Qwen2.5-0.5B-Instruct

0 commit comments

Comments
 (0)