The New Failure Modes in AI-Assisted Development
Amazon's AI coding incidents weren't caused by bad tools. They were caused by development processes that weren't redesigned to match the new risk profile.
Notes on AI architecture, agent systems, latency, and building production AI products.
Amazon's AI coding incidents weren't caused by bad tools. They were caused by development processes that weren't redesigned to match the new risk profile.
Voice AI sits at the extreme end of the latency spectrum. The pipeline, the UX, and the failure modes are fundamentally different from chat.
Most agent systems feel slow because they are architected to be slow. The model is rarely the bottleneck anymore, and measuring everything else is where the real gains are.
The gap between a compelling demo and a reliable production system is where most AI projects die. The problems are predictable, and the fixes are systematic.
Most teams asking for an agent really need a workflow. Autonomy has a cost, and the default should be simpler than you think.
Traditional testing gives you pass or fail. AI systems live in a probabilistic space where correctness is a spectrum and ground truth itself is unreliable.
RAG was a breakthrough, but its limitations are clear. When your knowledge base is complex, dynamic, or multi-modal, you need architectures that go further.