Skip to content

Conversation

@hidai25
Copy link

@hidai25 hidai25 commented Dec 5, 2025

Adding EvalView, an open-source agent testing framework :

  • Write YAML tests for LangGraph, CrewAI, OpenAI Assistants, Anthropic, custom APIs, etc.
  • Assert exact tool calls, LLM-as-judge output quality, cost & latency thresholds
  • Record real sessions → auto-generate hundreds of regression tests in minutes
  • No DB needed for local/CI runs

Already hit 4k+ views in a few hours on r/LangChain:
https://www.reddit.com/r/LangChain/comments/1pe3ivc/how_i_stopped_langgraph_agents_from_breaking_in/

Thanks for maintaining the list. This fits perfectly under Testing & Evaluation

Added EvalView section with details about the testing framework, including its features and links.
fixed the details part
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant