The techniques are broken down into two main streams, gradient and non-gradient approaches. Gradient approaches refers to fine-tuning the base LLM. While non-gradient approaches involves prompt engineering techniques which are delivered at inference.
Most notable are the inclusion of:
Hallucination mitigation in LLMs represents a multifaceted challenge addressed through a spectrum of innovative techniques.
Unlike traditional AI systems focused on limited tasks, LLMs have been exposed to vast amounts of online text data during training.
This allows LLMs to display impressive language fluency, it also means they are capable of:
This becomes hugely alarming when language generation capabilities are used for sensitive applications, such as:
The study includes very insightful taxonomy of hallucination mitigation techniques for LLMs; both gradient and non-gradient.
Gradient approaches include complex and opaque decoding strategies, knowledge graphs, fine-tuning strategies and more.
Non-gradient approaches include RAG, Self-Refinement and prompt tuning.
Notably the RAG approaches are segmented into four parts;
The power of prompt engineering to mitigate hallucination lies in defining:
The best defence against hallucination is not one single approach or method, but a combined approach defending against hallucination.
Seamlessly integrating numerous mitigation approaches, is the most important takeaway.
The factors which any organisation should keep in mind are:
Find the study here: https://arxiv.org/pdf/2401.01313
Previously published on Medium.