YOUR GEN AI SECRET WEAPON
AI + Reliability
At Leapfrogger, we've been diving deep into the world of multi-agent systems to harness the power of collaborative AI. TL;DR, it is a complex problem to crack. So it was with great anticipation that we listened to Ion Stoica at the Agentic AI Summit 2025 at Berkeley this past weekend. Ion, the co-founder of Databricks and Anyscale and a professor at UC Berkeley, discussed the failure modes of multi-agent systems using a clever combination of human review and AI-powered analysis. The results? A sobering reminder that many agentic systems fail more than half the time on complex tasks. That's a big red flag, and it underscores the urgent need to build more reliable AI.
The Fundamental Challenge of AI Reliability
Drawing from established systems engineering principles, Stoica defines reliability as "the ability of a system or component to function under stated conditions for a specified period of time." This translates to several critical properties for AI systems:
- Accuracy and Correctness: Getting the right answers, consistently.
- Consistency: Predictable behavior across similar situations.
- Predictability: Understanding how the system will perform.
- Robustness: Resilience in the face of unexpected or even malicious inputs.
- Safety: Protection against harmful or unintended consequences.
While demonstrating initial functionality is relatively easy, Stoica argues that achieving true reliability for production-ready AI requires a significant investment – often 10 to 50 times more effort. This additional effort focuses on:
- Rigorous testing and debugging: Finding and fixing those hidden errors.
- Compliance and governance: Ensuring responsible AI development and deployment.
- Observability and monitoring: Tracking performance and spotting anomalies.
- Error handling and recovery: Gracefully managing failures when they occur.
Why AI Systems Are Harder to Debug
Stoica pinpoints a fundamental challenge in debugging AI agents compared to traditional software: the specification problem. In traditional software, we have clear specifications, well-defined interfaces, and predictable outputs. AI systems, especially those powered by large language models (LLMs), throw a wrench in the works:
- Lack of Clear Specifications: LLMs rarely have precise specifications for expected behavior. Their outputs are often probabilistic and context-dependent.
- Error Detection Difficulty: Without clear specifications, it's hard to definitively determine when an error has occurred. The "correctness" of an LLM's output is often subjective and open to interpretation.
- Black Box Nature: Unlike traditional code, where developers can trace execution paths and inspect variable values, AI systems are often opaque. Understanding the internal workings of an LLM is extremely challenging.
- Output-Only Debugging: Developers are typically limited to observing the system's outputs and inferring the underlying problems. This makes debugging a much more indirect and challenging process.
Human-Centered Evaluation
Despite the rise of automated benchmarks, Stoica emphasizes that human evaluation is paramount for assessing AI reliability. He rightly points out that:
- Most successful AI applications today involve humans in the loop, directly or indirectly.
- Real-world performance demands evaluation with actual users and their specific needs.
- Human preferences are the ultimate measure of system utility and satisfaction.
- Traditional benchmarks often fail to capture the nuances of real-world usage.

LM Arena focuses on two key components of human preference:
- Substance: The factual accuracy, completeness, and logical consistency of the information.
- Style: The presentation and formatting of the AI's outputs, including visual elements, language tone, and overall user experience.
Stoica makes a crucial point: style isn't just superficial; it's fundamental to reliability. Just as UI/UX design impacts how effectively we interact with computers, the style of AI outputs determines how reliably we can process and act on the information.
Key findings from LM Arena highlight the impact of style on user preference:
- Answer Length: Longer answers are generally preferred, likely due to the perception of more detailed and comprehensive information.
- Formatting Elements: Markdown formatting, structured lists, headers, and bold text enhance readability and organization, positively influencing user preference.
- Sentiment and Tone: Positive framing and a professional tone are generally preferred over casual language.
- Emoji Usage: Surprisingly, a higher emoji count correlates negatively with user preference, suggesting that users may perceive excessive emoji usage as unprofessional or distracting.
MAST: Multi-Agent System Failure Taxonomy
Stoica's team conducted a systematic analysis of multi-agent system failures, using a combination of manual inspection and automated analysis with LLM judges. The research revealed that many agentic systems fail more than 50% of the time on challenging benchmarks, underscoring the urgent need for reliability improvements.

The MAST analysis categorizes these failures into three key areas, which we think are incredibly insightful for anyone considering multi-agent AI:
- Specification Issues: Think of this as a "garbage in, garbage out" problem, but for AI. If an agent doesn't clearly understand its task, gets confused about its role, or loses crucial information, it's destined to fail. It's like giving a team member vague instructions and expecting them to deliver a perfect result – it's just not going to happen.
- Inter-Agent Misalignment: This is where communication breakdowns cripple the system. Imagine a relay race where the baton gets dropped or the runners misunderstand each other. Information corruption, misinterpretations, and lost context can all derail the collaborative efforts of AI agents.
- Task Verification Issues: This is the equivalent of having a faulty quality control system. If your verification mechanisms aren't up to par, flawed implementations can slip through the cracks, leading to incorrect or unreliable outcomes.

So, what can we do right now?
Stoica rightly points out that achieving truly autonomous AI requires tackling the specification problem head-on, and borrowing proven reliability practices from established engineering disciplines. We couldn't agree more. Here's what that looks like in practice:
- Crystal-Clear Specifications: Define clear component specifications with well-defined interfaces and expected behaviors. Think of it as creating a detailed blueprint for each AI agent, leaving no room for ambiguity.
- Systematic Verification: Implement rigorous testing processes to automatically check systems against those specifications. This is about building a robust safety net to catch errors before they cause problems.
- Modular Design: Embrace modular designs that allow you to isolate, test, and replace individual components. This makes debugging and maintenance far easier, and it allows you to upgrade components without disrupting the entire system.
- Promote Reusability: Create components with clear specifications that can be easily integrated into new systems. This not only saves time and resources but also promotes consistency and reliability across your AI deployments.
In addition, the research community is actively working on these challenges. Stoica highlights ongoing efforts to develop open-source tools and platforms for analyzing agentic system traces, as well as expanding LM Arena to evaluate multimodal capabilities. This is a space to watch closely.
For CEOs, and product and technology leaders looking to leverage the power of GenAI, understanding these failure modes is paramount. By addressing the specification problem, adopting rigorous verification practices, and embracing modular designs, we can build more reliable and effective multi-agent AI systems that deliver real business value. Let's work together to navigate this exciting, but complex, landscape.
Ready to Build Something Amazing?
Let's discuss how we can help you avoid the common pitfalls and build products that people love and trust.
