Build an eval harness for 184 AI agent prompts with promptfoo

How to build an LLM-as-judge eval system that scores AI agent prompts on quality, identity, and safety.

March 30, 2026 · 9 min · Russell

Managing LLM context in a real application

How Claudriel manages LLM context in production: conversation trimming, turn budgets, model fallback, prompt caching, and per-turn token telemetry.

March 27, 2026 · 6 min · Russell