Abstract
A Go server is running. Everything looks fine. Then something breaks.
Latency jumps. CPU climbs. Memory starts behaving in ways you didn’t expect. And now you need answers fast.
In this workshop, we take a real Go application behind Caddy and turn it into something you can actually understand in production.
No heavy observability stack. No long setup phase. We focus on getting useful signals out of the system as quickly as possible: request visibility, latency tracking, resource usage, and runtime behavior.
We instrument, we observe, we debug real scenarios. The goal is not tooling for the sake of tooling: it’s understanding what your Go service is really doing when it runs under pressure.
By the end, you leave with a working, observable Go setup and a simple mental model to debug production issues without guessing.