Observability

5 practitioners working with Observability:

10 AI Agent Failure Modes: Why Agents Break in Production The documented ways AI agents fail: hallucination cascades, context overflow, tool calling errors, and 7 more. Diagnosis patterns and fixes for each.
Agent Observability How to implement distributed tracing, logging, and monitoring for AI agents using OpenTelemetry and purpose-built tools like Langfuse and Braintrust.
Debug Your RAG Pipeline Before Users Notice Monitor retrieval-augmented generation systems with OpenTelemetry tracing. Find whether bad answers come from retrieval, context, or generation.
LLM-as-Judge Evaluation Use LLMs to evaluate LLM outputs. Build reliable automated judges through critique shadowing and iterative calibration with domain experts.
Memory Attribution and Provenance Track where AI memories came from, when they were created, and how much to trust them

← All topics