Skip to content

Bump vLLM and TRL #665

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 2 commits into from
May 28, 2025
Merged

Bump vLLM and TRL #665

merged 2 commits into from
May 28, 2025

Conversation

lewtun
Copy link
Member

@lewtun lewtun commented May 28, 2025

I've tested the SFT and GRPO sample scripts still run :)

AIME24 eval for Llama 8B also fine:

Task Version Metric Value Stderr
all math_pass@1:1_samples 0.4000 ± 0.0910
math_pass@1:4_samples 0.4667 ± 0.0696
math_pass@1:8_samples 0.4667 ± 0.0656
math_pass@1:16_samples 0.4604 ± 0.0639
math_pass@1:32_samples 0.4594 ± 0.0634
math_pass@1:64_samples 0.4510 ± 0.0640
lighteval:aime24:0 2 math_pass@1:1_samples 0.4000 ± 0.0910
math_pass@1:4_samples 0.4667 ± 0.0696
math_pass@1:8_samples 0.4667 ± 0.0656
math_pass@1:16_samples 0.4604 ± 0.0639
math_pass@1:32_samples 0.4594 ± 0.0634
math_pass@1:64_samples 0.4510 ± 0.0640

Ran with

export VLLM_WORKER_MULTIPROC_METHOD=spawn
NUM_GPUS=8
MODEL=deepseek-ai/DeepSeek-R1-Distill-Llama-8B
MODEL_ARGS="model_name=$MODEL,dtype=bfloat16,data_parallel_size=$NUM_GPUS,max_model_length=32768,gpu_memory_utilization=0.8,generation_parameters={max_new_tokens:32768,temperature:0.6,top_p:0.95}"
TASK=aime24
OUTPUT_DIR=data/evals/$MODEL

lighteval vllm $MODEL_ARGS "lighteval|$TASK|0|0" \
    --use-chat-template \
    --output-dir $OUTPUT_DIR

@lewtun lewtun requested a review from edbeeching May 28, 2025 06:48
@lewtun lewtun merged commit b806e10 into main May 28, 2025
1 check passed
@lewtun lewtun deleted the bump-deps-0 branch May 28, 2025 11:47
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants