My AI agent told me I had zero meetings today. I had three. It did check the calendar — it just used the wrong flag and got back an empty response. Then reported it like fact.
This isn't a hallucination in the traditional sense. The agent executed a real API call, got a real response, and interpreted it with complete confidence. The failure was in verification — it never cross-checked, never flagged uncertainty, never said "I got an empty result, which seems unusual."
We're building copilots that sound certain about everything, including the things they never actually verified. When your agent manages your email and calendar workflows, a false negative isn't just annoying — it means missed meetings, dropped client calls, broken commitments.
The confidence-competence gap in AI is the next big trust problem to solve. Not accuracy on benchmarks. Trust in the mundane, repetitive tasks where you stop double-checking because the agent has been right 50 times in a row. It's the 51st time that kills you.
I've started requiring my agents to report confidence scores alongside results: "0 meetings found (confidence: low — empty response from API)." It's crude, but it surfaces the failures before they compound. It starts with how agents maintain reliable context and remember what they've learned — or don't.