Over the last 24 hours, I've run into three completely different situations that all point to the same, unnerving conclusion: our collective trust in AI is dangerously high, and our processes are terrifyingly immature. We're handing people a powerful tool, but we're forgetting the safety manual and, most importantly, the guard.
AI - and Large Language Models (LLMs) in particular - are like a chainsaw. In the hands of a skilled operator who respects the tool, it can clear a forest. In the hands of a novice, it can take off a leg. Right now, a lot of businesses are gleefully revving up the engine with no guard in sight.
Exhibit A: The âHyperpersonalizedâ Fail
It started this morning with a cold email.
David, I want to introduce you to Newstool: a single email each morning with everything your competitors are doing...
Can I send a sample comparing Cortado Group against Menzies Aviation and Alliance Ground International?
I had to laugh. The AI-powered tool this salesperson used to find our âtop competitorsâ pulled two names that are so wildly incorrect, it's clear they have no idea what our company does. Their attempt at âhyperpersonalizationâ immediately exposed their process as a hollow gimmick.
This is a classic case of Garbage In, Garbage Out.
The AI didnât fail - the data it was fed failed. The system gobbled up bad data and confidently spat out a personalized error.
I replied immediately, asking how they identified those companies.
Thirteen hours later? Silence. The window for a meaningful conversation slammed shut long ago.
Research from institutions like Harvard Business Review has shown that firms that try to contact potential customers within an hour of receiving a query are nearly seven times as likely to have a meaningful conversation as those that waited even an hour longer.
After 13 hours, my interest is gone. Their over-reliance on a flawed automated process didnât just fail to start a conversation - it actively prevented one.
Exhibit B: The Internal âBlack Boxâ Dilemma
The second instance came from an internal discussion with our innovation team. We were debating the merits of different data enrichment tools, and a key theme emerged: a deep-seated distrust of âblack boxâ AI.
One of my team members, Mitch, pointed out that a tool we were considering was just âtoo black box.â He recounted a near-miss with a major client, where the AIâs opaque reasoning almost caused a significant problem. We simply couldnât explain how the AI arrived at its conclusions.
This hits on two critical principles of data quality:
- AI Explainability: The AI âshoves all the stuff in a box and then comes up with correlations that you might not understand.â
If you canât understand the logic, you canât trust the output. Youâre just blindly following a machineâs recommendation. - Data Provenance: Where did this information come from?
If you canât trace data back to its source, you canât verify its accuracy or reliability. Itâs the data equivalent of âsome guy told me.â
Without transparency and traceability, youâre not using a tool - youâre gambling.
Exhibit C: The Hallucination Nightmare
The final, and most alarming, piece of the puzzle came from two separate conversations.
First, one of our sharpest consultants, Jessica, expressed her mounting frustration:
I'm still dealing with ChatGPT making things up, and it's extremely frustrating.
This isnât a minor bug; itâs a fundamental flaw we call hallucination.
The AI isnât just wrong - itâs confidently and creatively wrong.
Her frustration was put into terrifying context when our RevOps practice lead, Robert, brought a recent news story to my attention.
A team at Deloitte had used an LLM to help synthesize research for a report for the Australian government. The AI did what it sometimes does: it hallucinated - fabricating completely false information, including citations to non-existent academic research.
The result was a public scandal that became a âhuge black markâ on the use of LLMs in consulting.
The tool that was meant to create efficiency instead created a crisis of credibility.
Itâs Time to Build the Guardrails
These three stories - a failed sales pitch, an internal debate on opaque tools, and a real-world hallucination disaster - are symptoms of the same disease.
Weâre treating AI like a magic wand instead of the powerful, dangerous, and often flawed tool it is.
Saying âeveryone should just use ChatGPTâ is like saying âeveryone should just use a chainsaw.â
Itâs reckless.
The solution isnât to abandon the tool - itâs to build the guardrails and implement rigorous processes.
This is what true AI adoption looks like. Itâs not about access; itâs about architecture.
Hereâs what that looks like in practice:
-
Automated Verification
Stop relying on humans to manually fact-check everything.
Build programmatic validation layers into your process.
For our interview analysis tool, weâre implementing evidence thresholds:If the evidence for this conclusion is greater than 70%, give me a âyes.â
-
Source Grounding
Require the AI to cite its sources for any factual claim.
If it canât show you where it got the information, the information is suspect by default. - Structured Prompts
Move away from simple chat interfaces for critical tasks.
Use structured schemas that force the model to separate facts from assumptions. - A âRefuse When Unsureâ Rule
The most important guardrail of all.
Program your systems to prefer an explicit âI donât have enough informationâ over a confident but fabricated guess.
From âPrompt-and-Prayâ to Proof and Process
We need to shift our mindset from âprompt-and-prayâ to one of systematic, automated validation.
The real work of AI innovation isnât just in using the tool - itâs in building the safety systems that make it reliable, trustworthy, and ready for the enterprise.
Otherwise, weâre just waiting for the accident to happen.