If your company is like most, generative AI (GenAI) is already at work. Probably, you’ve stood up a few use cases. Maybe (as we have done at PwC) you’ve already scaled up some. And, most likely, your enterprise software now has embedded GenAI capabilities, which your people are using every day. But as GenAI becomes more critical to daily operations, a common question persists: Can I trust it?
Trust in GenAI requires all the traditional drivers of trust in tech: governance, security, compliance and privacy. As with traditional AI, you also need to mitigate bias. But GenAI also adds a new risk, “hallucinations” — outputs that seem plausible but have no basis in reality.
Traditional AI sometimes hallucinates too, but more rarely, and it’s usually operated by specialists who can catch these errors. With GenAI, hallucinations can — in some specific instances — pose a risk, especially for GenAI users who don't have deep AI experience.
The good news is you can manage this risk — and do so in ways that speed up your AI initiatives, not slow them down. Here’s what you need to know about hallucinations and how a robust Responsible AI approach can mitigate their risk.
GenAI is designed to provide outputs most likely to be correct, based on the data and prompts provided. It will often “try its best” even when it has overly complex or incomplete data. This creativity helps give GenAI the flexibility to accomplish many different tasks, including some that only a human could previously do. But, GenAI’s creativity can sometimes go too far— most commonly when one or more of these situations occur:
Mitigating the risks of hallucinations has three main components, which are part of Responsible AI:
To keep down the number of hallucinations, choose the right GenAI solution for the right use case — and be sure that it has adequate, reliable data. Identify and call out those use cases where data may be not good enough. You still may choose to proceed with these use cases, if their value is high enough, but they should be flagged for extra oversight.
Prepare your people with upskilling and templates to give GenAI suitable prompts. Set up guardrails to help people use GenAI only for its intended tasks.
Technology can help catch hallucinations. Retrieval-augmented generation (RAG) can enable GenAI to “double-check” an authoritative knowledge base, and application programming interfaces (APIs) can connect to approved content. Teach your people how to verify GenAI’s outputs, create channels to report suspect results, and establish a risk-based system to review these outputs. Finally, have people train GenAI, based on its mistakes, so it will make fewer going forward.
Even though hallucinations are a moving target, you can manage the risk. If well-deployed, these Responsible AI measures won’t be onerous and they won’t slow you down. On the contrary, AI initiatives can proceed more rapidly when they cause fewer mistakes and don’t require costly fixes and do-overs. Here are measures you can take:
We don't expect hallucinations to go away. We do expect them to change in nature, while companies’ approach to them becomes more sophisticated.
As GenAI models improve, they will be better able to recognize nuances in language, track long-term dependencies in text, understand the intent and context of a user’s request and address mathematical challenges. They will become ever more multimodal, incorporating multiple forms of data such as text, images and audio. All these improvements (and others too) could help reduce or eliminate some hallucinations that are common today.
But as people use GenAI for more tasks — in medical diagnoses, market and economic forecasts, fundamental scientific research and more — new kinds of hallucinations will likely emerge and the cycle will continue.
Hallucinations are, in short, an inevitable part of GenAI’s “creative” process: Like a person, as GenAI tries to innovate and find new insights, it will err sometimes. Companies should be ready to catch these mistakes, so GenAI too can “fail fast” and do better next time.
Lead with trust to drive outcomes and transform the future of your business.