Woolworths AI Glitch Highlights Risks of Hybrid Legacy-LLM Retail Systems
Key Takeaways
- Woolworths' AI assistant, Olive, recently drew scrutiny after providing bizarre personal anecdotes and inaccurate pricing to customers.
- The incident underscores the technical challenges of integrating modern large language models with legacy decision-tree scripts and real-time inventory databases.
Mentioned
Key Intelligence
Key Facts
- 1Olive is Woolworths' AI assistant powered by a Large Language Model (LLM).
- 2Bizarre responses about a 'mother' were traced to legacy pre-written scripts from several years ago.
- 3Pricing errors occurred because the LLM was not effectively 'grounded' in real-time inventory data.
- 4Woolworths has since removed the problematic scripting following customer feedback.
- 5The incident mirrors a 2022 Air Canada case where a chatbot's misinformation led to a legal ruling against the airline.
| Company | ||
|---|---|---|
| Woolworths | Legacy script interference & pricing errors | Problematic scripts removed |
| Air Canada | Hallucinated bereavement fare policy | Legally forced to refund passenger |
| DPD | Rogue poetry and inappropriate language | Chatbot functionality disabled |
Analysis
The recent erratic behavior of Woolworths’ AI assistant, Olive, serves as a stark reminder that the "move fast and break things" ethos of Silicon Valley can have embarrassing, and potentially costly, consequences when applied to the high-stakes world of retail customer service. Australian shoppers interacting with Olive were met with responses that veered from the helpful to the surreal, with the bot claiming to have a "mother" and providing incorrect pricing for staple goods. While these interactions might seem like harmless digital quirks, they expose a deeper structural flaw in how major corporations are currently deploying generative artificial intelligence: the messy collision of cutting-edge Large Language Models (LLMs) with aging legacy infrastructure.
According to Woolworths, the references to Olive’s "mother" were not a case of the AI developing a persona or "hallucinating" in the traditional sense. Instead, they were the result of old decision-tree scripts—some dating back several years—that remained buried within the system’s logic. When a user entered data that the system interpreted as a birthdate, it triggered a "fun fact" script from a previous era of chatbot technology. This highlights a significant challenge for enterprise-level AI rollouts: technical debt. Many retailers are not building AI from scratch but are layering LLMs over existing customer service frameworks. When these layers aren't perfectly synchronized, the result is a "Frankenstein’s monster" of a bot that oscillates between sophisticated language generation and rigid, outdated pre-programmed responses.
According to Woolworths, the references to Olive’s "mother" were not a case of the AI developing a persona or "hallucinating" in the traditional sense.
The pricing errors reported by users point to an even more critical failure known as a lack of "grounding." LLMs are probabilistic engines; they predict the next most likely word in a sequence based on training data, not real-time facts. For a retail AI to be useful, it must be tethered to a live database of inventory and pricing—a process often achieved through Retrieval-Augmented Generation (RAG). If this connection is weak or intermittent, the AI will rely on its training data, which may be months or years out of date, or it will simply guess based on patterns. In a retail environment where price sensitivity is at an all-time high due to inflation, providing a customer with an incorrect price isn't just a technical glitch; it’s a breach of trust that can lead to accusations of price gouging or bait-and-switch tactics.
What to Watch
This incident does not exist in a vacuum. The retail and service industries have seen a string of high-profile AI failures that are beginning to shape the legal and regulatory landscape. In 2022, Air Canada was held liable for a chatbot that invented its own bereavement fare policy, misleading a passenger named Jake Moffatt. The tribunal ruled that the airline was responsible for the information provided by its bot, regardless of whether it was "hallucinated." Similarly, the courier service DPD had to disable its AI after it was goaded into writing poems about how bad the company was. These precedents suggest that the "it’s just a beta" excuse will no longer hold water in court or in the court of public opinion.
For e-commerce leaders, the Woolworths case is a signal that the honeymoon phase of AI experimentation is ending. The focus must now shift from "can we build it?" to "can we govern it?" This requires rigorous testing of edge cases, the implementation of robust guardrails to prevent legacy scripts from interfering with LLM outputs, and ensuring that any AI-facing customer data is grounded in a "single source of truth" for pricing and inventory. As AI becomes the primary interface for the customer journey, the cost of a "rambling" bot is no longer just a PR headache—it is a direct threat to the bottom line and brand equity. Retailers must prioritize technical hygiene over rapid deployment to avoid the reputational damage that comes when their digital representatives go rogue.
Sources
Sources
Based on 2 source articles- (in)Woolworths AI agent rambled about its mother. Its a sign of deeper problems with the tech rollout - The TribuneFeb 28, 2026
- List.metadata.agency (in)Woolworths’ AI agent rambled about its ‘mother’. It’s a sign of deeper problems with the tech rolloutFeb 28, 2026