{
  "@context": "https://schema.org",
  "@type": "QAPage",
  "canonical": "https://ireadcustomer.com/en/blog/air-canada-lost-in-court-to-its-own-ai-what-custom-guardrails-will-save-you",
  "markdown_url": "https://ireadcustomer.com/en/blog/air-canada-lost-in-court-to-its-own-ai-what-custom-guardrails-will-save-you.md",
  "title": "Air Canada Lost in Court to Its Own AI — What Custom Guardrails Will Save You",
  "locale": "en",
  "description": "Air Canada was forced to pay damages after its customer service chatbot invented a fake refund policy. Here is how to build the custom AI guardrails that prevent your business from facing the same legal disaster.",
  "quick_answer": "A tribunal ruled Air Canada liable for damages after its customer service AI chatbot hallucinated a fake refund policy, establishing that companies are legally responsible for their chatbots. Businesses must implement retrieval grounding and human-in-the-loop guardrails to avoid similar legal exposure.",
  "summary": "In November 2022, Jake Moffatt was scrambling to book a last-minute flight after his grandmother passed away. He visited the Air Canada website and asked the customer support chatbot if the airline offered bereavement rates. The chatbot confidently replied that Moffatt could book a regular ticket immediately and request a refund for the price difference within 90 days. Relieved, he used his credit card to book the flight on the spot. But the chatbot lied. Air Canada's actual bereavement policy strictly states that discounts cannot be applied retroactively. When Moffatt asked for his money back",
  "faq": [
    {
      "question": "What happened in the Moffatt v. Air Canada chatbot lawsuit?",
      "answer": "A customer sued Air Canada after its website chatbot invented a fake bereavement refund policy. The tribunal ruled against the airline, stating that a company is completely legally liable for the information provided by the AI tools hosted on its website."
    },
    {
      "question": "Why do customer service chatbots hallucinate fake policies?",
      "answer": "Standard off-the-shelf AI models are designed to be conversational and helpful, not purely factual. They predict the most logical next word based on broad training data. If they don't know your specific internal policy, they will confidently guess or invent one to satisfy the user's question."
    },
    {
      "question": "What is retrieval grounding in AI chatbots?",
      "answer": "Retrieval grounding is a technical guardrail where the AI is disconnected from its general internet knowledge and restricted to only reading and answering from a specific set of verified corporate documents, like official PDF return policies or pricing sheets."
    },
    {
      "question": "How can businesses legally protect themselves when using AI for customer support?",
      "answer": "Businesses must implement strict guardrails: use retrieval grounding to force citations from official policies, program verifiable refusals so the bot knows when to say no, and require a human-in-the-loop to approve any actions involving money, contracts, or refunds."
    }
  ],
  "tags": [
    "ai hallucination",
    "chatbot liability",
    "retrieval augmented generation",
    "ai guardrails",
    "customer service automation"
  ],
  "categories": [],
  "source_urls": [],
  "datePublished": "2026-05-07T01:24:01.051Z",
  "dateModified": "2026-05-07T01:24:01.061Z",
  "author": "iReadCustomer Team"
}