Add RolloutGateway to enable native CLI agent training#1757
Open
Add RolloutGateway to enable native CLI agent training#1757
RolloutGateway to enable native CLI agent training#1757Conversation
samsja
reviewed
Feb 16, 2026
samsja
reviewed
Feb 16, 2026
samsja
reviewed
Feb 16, 2026
samsja
reviewed
Feb 16, 2026
samsja
reviewed
Feb 16, 2026
samsja
reviewed
Feb 16, 2026
12a3171 to
db0d01d
Compare
13 tasks
CliAgentEnv trainingRolloutGateway to enable native CLI agent training
There was a problem hiding this comment.
Cursor Bugbot has reviewed your changes and found 1 potential issue.
Bugbot Autofix is OFF. To automatically fix reported issues with Cloud Agents, enable Autofix in the Cursor dashboard.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Adds a rollout gateway to the vLLM inference server, enabling server-side multi-turn rollout execution.
depends on PrimeIntellect-ai/verifiers#954
Rollout gateway (rollout_gateway.py, 542 lines — new file):
Inference server (server.py):
Config (config.py, inference/config.py):
Note
Medium Risk
Adds new FastAPI endpoints and per-rollout state management inside the vLLM inference server, plus tweaks DP/api-server scaling behavior; issues here could break inference serving or rollout collection in production.
Overview
Introduces a new inference-side rollout gateway (
/v1/rollouts/...) that lets clients register multi-turn rollouts, route requests to a data-parallel rank, and fetch token-level trajectories (ids/masks/logprobs), with optional verbose turn logging.Updates inference config to add
log_rollout_gateway_turnsandauto_scale_api_servers(gating the existing dp→api_server_countauto-scaling), and mounts/initializes the gateway in the vLLM server (disabled whenapi_server_count>1). Also updates orchestrator metric logging to pass an explicitstep.Written by Cursor Bugbot for commit 27806a4. This will update automatically on new commits. Configure here.