Observability

Building the future of agents with Claude

Building the future of agents with Claude

Experts from Anthropic discuss the evolution of the Claude Developer Platform, the philosophy of "unhobbling" models with tools rather than restrictive scaffolding, and the future of building sophisticated, autonomous AI agents with features like the Claude Agent SDK, advanced context management, and persistent memory.

Why You Should Care About Observability in LLM Workflows

Why You Should Care About Observability in LLM Workflows

An inside look at AlwaysCool.ai's journey from simple GPT wrappers to a production-ready agentic infrastructure. This talk covers the evolution from synchronous tools to asynchronous, multi-step flows orchestrated by LangGraph, the critical role of OpenTelemetry for compliance and observability, and the architectural patterns of using FastAPI to serve centralized AI agents.

Evaluating AI Agents: Why It Matters and How We Do It

Evaluating AI Agents: Why It Matters and How We Do It

Annie Condon and Jeff Groom from Acre Security detail their practical approach to robustly evaluating non-deterministic AI agents. They share their philosophy that evaluations are critical for quality, introduce their "X-ray machine" analogy for observability, and walk through their evaluation stack, including versioning strategies and the use of tools like Logfire for tracing and Confident AI (Deep Evals) for systematic metric tracking.

Building Multi-Player AI Systems (and why it’s SO hard)

Building Multi-Player AI Systems (and why it’s SO hard)

MeshAgent introduces a multiplayer AI paradigm, shifting from single-user systems to collaborative 'Rooms' where teams of humans and agents can work together with shared context. This talk explores the platform's architecture, developer tools, and its approach to solving real-world collaborative tasks.

Reliability Engineering Mindset • Alex Ewerlöf & Charity Majors • GOTO 2025

Reliability Engineering Mindset • Alex Ewerlöf & Charity Majors • GOTO 2025

Alex Ewerlöf, author of "Reliability Engineering Mindset," discusses the significant gap between Google's idealized SRE practices and the resource-constrained reality of most companies. The conversation focuses on making Service Level Objectives (SLOs) practical by tying Service Level Indicators (SLIs) directly to business impact, using them as a data-driven communication tool to negotiate reliability costs, and moving from a "best practice" to a "fit practice" mindset.

From Spikes to Stories: AI-Augmented Troubleshooting in the Network Wild // Shraddha Yeole

From Spikes to Stories: AI-Augmented Troubleshooting in the Network Wild // Shraddha Yeole

Shraddha Yeole from Cisco ThousandEyes explains how they are transforming network observability by moving from complex dashboards to AI-augmented storytelling. The session details their use of an LLM-powered agent to interpret vast telemetry data, accelerate fault isolation, and improve MTTR, covering the technical architecture, advanced prompt engineering techniques, evaluation strategies, and key challenges.