My AI agent cost me $400 overnight, so I built pytest for agents and open-sourced it
github.comΒ·3hΒ·
Discuss: DEV
🚨Lighthouse CI
Preview
Report Post

EvalView β€” Pytest-style Testing for AI Agents

The open-source testing framework for LangGraph, CrewAI, OpenAI Assistants, and Anthropic Claude agents. Write tests in YAML, catch regressions in CI, and ship with confidence.

EvalView is pytest for AI agentsβ€”write readable test cases, run them in CI/CD, and block deploys when behavior, cost, or latency regresses.


What is EvalView?

EvalView is a testing framework for AI agents.

It lets you:

  • πŸ§ͺ Write tests in YAML that describe inputs, expected tools, and acceptance thresholds
  • πŸ” Turn real conversations into regression suites (record β†’ generate tests β†’ re-run on every change)
  • 🚦 Gate deployments in CI on behavior, tool calls, cost, and latency
  • 🧩 Plug into **LangGraph, CrewAI, OpenAI Assistants, …

Similar Posts

Loading similar posts...