Karthik Mahalingam writing at LatentMesh
How AI systems fail.
How to evaluate them.
How to govern them in production.
Most AI governance advice skips the engineering. These essays don't.
Series: Reliable Agent Systems · Ten essays
Start here
New to LatentMesh? Start with the core essay series on how agent systems fail, how to evaluate them in the real world, and how governance becomes operational.
Agent Failures Are Distributed Systems Failures
Why agent failures are usually system failures, not just model failures.
The Eval Gap
Why staging success does not predict production reliability.
Controls Are Not Guardrails
Why runtime safety layers are not enough without evidence, ownership, and testing.
Latest essays
Recent essays on agent reliability, evaluation, safety, and governance.
What I write about
Agent failure as system failure
How agent systems break, drift, and misbehave once they leave the demo.
Evaluation that catches real failures
Practical evals for regression, reliability, and operational trust.
Safety as systems engineering
Guardrails, control surfaces, observability, containment, and human review.
Compliance that becomes architecture
How legal and policy obligations turn into workflows, controls, and evidence.
About
I write about AI systems from the perspective of someone who has spent nearly two decades building large-scale distributed systems. My focus is on how agents behave in production, how evaluations catch real failures, and how safety, privacy, and compliance become operational reality.
More about me →