Skip to content
#

evaluation-suite

Here is 1 public repository matching this topic...

Extensible benchmarking suite for evaluating AI coding agents on web search tasks. Compare native search vs MCP servers (You.com, expanding) across multiple agents (Claude Code, Gemini, Droid, Codex, expanding) with automated Docker workflows and statistical analysis.

  • Updated Jan 30, 2026
  • TypeScript

Improve this page

Add a description, image, and links to the evaluation-suite topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the evaluation-suite topic, visit your repo's landing page and select "manage topics."

Learn more