Skip to content

feat: streaming SSE for benchmark LLM calls with direct endpoint support#129

Merged
solderzzc merged 1 commit intodevelopfrom
feat/streaming-benchmark-env-vars
Mar 7, 2026
Merged

feat: streaming SSE for benchmark LLM calls with direct endpoint support#129
solderzzc merged 1 commit intodevelopfrom
feat/streaming-benchmark-env-vars

Conversation

@solderzzc
Copy link
Member

  • Rewrite llmCall() to use stream:true with SSE parsing and idle timeout
  • Support direct llama-server (AEGIS_LLM_URL) and cloud provider (AEGIS_LLM_BASE_URL + API key) endpoints, bypassing gateway
  • Handle reasoning_content from thinking models (Qwen3.5)
  • Add progress logging every 100 tokens
  • Update banner and healthcheck to show actual LLM endpoint

- Rewrite llmCall() to use stream:true with SSE parsing and idle timeout
- Support direct llama-server (AEGIS_LLM_URL) and cloud provider
  (AEGIS_LLM_BASE_URL + API key) endpoints, bypassing gateway
- Handle reasoning_content from thinking models (Qwen3.5)
- Add progress logging every 100 tokens
- Update banner and healthcheck to show actual LLM endpoint
@solderzzc solderzzc merged commit 05344e8 into develop Mar 7, 2026
1 check passed
@solderzzc solderzzc deleted the feat/streaming-benchmark-env-vars branch March 7, 2026 15:22
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant