Terrarium
Evals that actually simulate real deployment — stateful, multi-turn, alive
Terrarium is a multi-turn evaluation and optimization engine for LLM agents built by evolvent-ai. Unlike static benchmark suites that measure agents against fixed input-output pairs, Terrarium creates persistent, stateful "living environments" — simulated deployment contexts where agents operate over extended sessions, accumulate state, use tools, and interact with simulated external systems. You evaluate agents the way you'd test a car: by driving it, not by measuring its doors. The system supports configurable environment complexity, including simulated databases, APIs, file systems, and user personas. Agents are scored not just on final outputs but on trajectory quality — how efficiently they reached the answer, how often they hallucinated intermediate steps, and how well they recovered from dead ends. The engine also supports continuous optimization loops where poor-performing trajectories trigger automatic prompt refinement. With 17 stars and created April 14, Terrarium is extremely new. But it's addressing a genuine gap: the disconnect between how agents perform on static benchmarks versus how they behave in production. As enterprise AI deployments scale, the need for realistic pre-production evaluation is becoming critical.
Panel Reviews
The Builder
Developer Perspective
“Static evals are lying to us constantly — agents that ace benchmarks fall apart in production because benchmarks don't have state, side effects, or accumulated context. Terrarium's living environments model is the right approach to catching real failure modes before deployment.”
The Skeptic
Reality Check
“Building a realistic simulation of your production environment is often harder than just running the agent in staging. The value proposition assumes your eval environment is meaningfully closer to production than your existing test suite — which is a big assumption for complex deployments.”
The Futurist
Big Picture
“The eval-optimize loop is the missing piece in most AI agent development workflows. Tools that can automatically identify weak trajectories and suggest improvements will become as fundamental as unit tests. Terrarium is early, but the category is inevitable.”
The Creator
Content & Design
“This is deeply technical infrastructure that won't affect my daily workflow. The people who need this know they need it — but for most creators building with AI tools, static evals are already more than they use.”
Community Sentiment
“Stateful evals vs. static benchmarks”
“Simulated environments for pre-production agent testing”
“Living environments and trajectory quality scoring”