ENGINEERING

Why We Built Anti-Hallucination Guardrails Into Our AI

Andres MuguiraFebruary 18, 20267 min read
AISafetyEngineering
← Back to Blog
Summarize with AI

AI in a CRM Is Not a Toy

When AI hallucinates in a chatbot conversation, the stakes are low. Someone gets a wrong answer, laughs it off, and tries again. When AI hallucinates in a CRM, the stakes are completely different. A hallucinated email gets sent to a real client. A fabricated deal value corrupts your forecast. A made-up contact detail wastes hours of follow-up on a phone number that does not exist.

We learned this the hard way during our alpha. An early version of our AI assistant confidently told a user that their prospect "had previously expressed interest in the enterprise plan during a call on March 12th." There was no call on March 12th. The AI had invented context that sounded plausible but was entirely fictional. The user drafted a follow-up email referencing that nonexistent conversation. Fortunately, they caught it before sending. But it exposed a fundamental problem: in a CRM, AI confidence without accuracy is dangerous.

That incident led us to build a multi-layered anti-hallucination system. This post explains each layer, why it exists, and where we still have limitations.

In a CRM, AI confidence without accuracy is not just unhelpful -- it is actively harmful. A hallucinated fact in a client email can destroy months of relationship building.
The AI chat assistant executes tool calls with reasoning steps visible - every action is grounded in real CRM data

Layer 1: Draft Confirmations Before Execution

The most important guardrail is also the simplest: the AI never executes actions autonomously. It drafts them. Every action the AI proposes -- sending an email, updating a deal value, creating a task, modifying a contact record -- is presented as a draft that requires human confirmation before execution.

This might sound like it slows things down, and it does, slightly. But the tradeoff is worth it. The confirmation step takes 2-3 seconds for a human to review. An incorrect autonomous action can take hours to undo and may cause irreversible damage to a client relationship.

The draft confirmation UI is designed for speed. The proposed action is displayed in a clear, scannable format with the key changes highlighted. One click to approve, one click to reject, or inline editing if the draft is 90% correct but needs a small adjustment. Most users develop a rhythm: glance, approve, move on. The friction is minimal, but it catches the 5-10% of cases where the AI gets something wrong.

What draft confirmations catch

Layer 2: Tool-Level Validation

Under the hood, our AI assistant uses a tool-calling architecture. When the AI decides it needs to take an action, it calls a specific tool (send_email, update_deal, create_task) with structured parameters. Each tool has its own validation layer that checks the parameters before the action is presented to the user.

For example, the send_email tool validates:

The update_deal tool validates that proposed values fall within reasonable ranges based on historical data. If your average deal size is $5,000 and the AI proposes updating a deal to $500,000, the validation layer flags it for explicit confirmation with a warning.

These validations are not AI-based. They are deterministic rules. The AI generates the intent, but hard-coded logic validates the parameters. This separation is important because it means the validation layer cannot be fooled by a persuasive but incorrect AI response.

Layer 3: Context Boundaries

Hallucinations often happen when AI models operate beyond the boundaries of their available context. If you ask the AI "What did Sarah say in our last meeting?" and there are no meeting notes for Sarah, the model has two options: admit it does not know, or fabricate a plausible answer. Without guardrails, models tend toward fabrication because they are trained to be helpful.

Our context boundary system explicitly defines what data the AI can and cannot access for each query:

When the AI attempts to reference data outside its current context boundary, the system returns a structured "no data available" response instead of allowing the model to guess. The AI then surfaces this to the user: "I do not have meeting notes for Sarah Chen. Would you like to add notes from your last conversation?"

This is less impressive than a confident (but wrong) answer. It is also infinitely more useful.

Layer 4: Source Attribution

Every factual claim the AI makes in SalesSheet includes a source attribution. When the AI says "Your deal with Acme Corp has been in the Proposal stage for 14 days," that statement links back to the deal record and stage change timestamp that generated it. When the AI says "Sarah's last email mentioned budget concerns," it links to the specific email.

Source attribution serves two purposes. First, it lets users verify claims instantly by clicking through to the source data. Second, it constrains the AI's behavior during generation. The model is instructed to only make claims it can attribute to specific data sources. If it cannot cite a source, it cannot make the claim.

This does not eliminate hallucinations entirely. Models can still misinterpret source data or draw incorrect conclusions from valid data. But it dramatically reduces the most dangerous type of hallucination: fabricated facts with no basis in reality.

Layer 5: Confidence Signals

Not all AI outputs carry the same level of certainty. Summarizing an email thread is high-confidence work -- the data is right there. Predicting whether a deal will close is inherently uncertain. We built a confidence signal system that communicates this distinction to users.

High-confidence outputs (summaries, data lookups, calculations) are presented as statements. Low-confidence outputs (predictions, suggestions, inferences) are presented with hedging language and explicit uncertainty markers. The UI uses subtle visual cues: a solid icon for high-confidence responses, a dashed outline for uncertain ones.

We considered hiding low-confidence outputs entirely, but decided against it. Users find uncertain suggestions useful as starting points for their own thinking. The key is making the uncertainty visible so users calibrate their trust appropriately.

What We Still Get Wrong

Honesty about limitations matters more than marketing claims about accuracy. Here is what our anti-hallucination system does not catch:

We track hallucination reports from users and review them weekly. Every confirmed hallucination becomes a test case in our validation suite. The system gets better over time, but it will never be perfect. The goal is not perfection -- it is making failures visible, recoverable, and rare.

The goal is not zero hallucinations. It is zero undetected hallucinations. Every error should be visible, every action should be reversible, and the user should always have the final say.

The Trust Equation

AI trust in a CRM follows a simple equation: trust equals accuracy multiplied by transparency. High accuracy with low transparency produces anxiety -- users know the AI is usually right but cannot tell when it is wrong. Low accuracy with high transparency produces annoyance -- users see too many warnings and caveats. High accuracy with high transparency produces confidence -- users trust the AI because they can verify it and they rarely need to.

Our anti-hallucination system is designed to maximize both sides of that equation. The validation layers increase accuracy. The draft confirmations, source attributions, and confidence signals increase transparency. Together, they create an AI assistant that users actually rely on for real sales work -- not because they blindly trust it, but because they have the tools to verify it quickly and the confidence that it will not take irreversible actions without their approval.

If you are evaluating AI-powered CRMs, ask the vendor one question: "What happens when your AI is wrong?" If they cannot give you a specific, detailed answer, that tells you everything you need to know about how much thought they have put into the problem.

Try SalesSheet Free

No credit card required. Start selling smarter today.

Start Free Trial