When Your "New Teammate" Hallucinates: Observability as the Safety Net for AI Code

As AI copilots and coding agents generate more of the code we ship, we can deliver software faster than we can understand it, leading to fragile behavior, missed edge cases, and failures that are hard to explain.

This talk’s guiding principle is simple: the faster you generate code, the more you need telemetry. We’ll connect observability fundamentals (signals, context propagation, debugging workflows) to the realities of AI-generated code, using OpenTelemetry as a vendor-neutral foundation for traces, metrics, and logs that reveal runtime truth.

You’ll see three practical approaches: prompt AI tools to emit instrumentation by default (using a reusable template), use auto-instrumentation for broad, fast coverage, and add manual spans/attributes where domain context matters most. To make this stick, we’ll introduce a simple team practice: require “observability present” in code review for AI-generated changes.

Speaker

florian-mair

Florian Mair

 
Florian is a passionate software engineer & architect with 10+ years of experience in Go, Java and JS. He has built full-stack applications and data processing pipelines that handle millions of ...