<img src="https://secure.item0self.com/192096.png" alt="" style="display:none;">

Large Language Model Hallucination Mitigation Techniques

This recently released study is a comprehensive survey of 32+ mitigation techniques to address hallucination.


Introduction

The techniques are broken down into two main streams, gradient and non-gradient approaches. Gradient approaches refers to fine-tuning the base LLM. While non-gradient approaches involves prompt engineering techniques which are delivered at inference.

1*oiR_cRtoJa_VlIopzcDO8g-2-2

Most notable are the inclusion of:

  1. Retrieval-Augmented Generation (RAG)
  2. Knowledge Retrieval
  3. CoNLI
  4. CoVe

Hallucination mitigation in LLMs represents a multifaceted challenge addressed through a spectrum of innovative techniques.

Unlike traditional AI systems focused on limited tasks, LLMs have been exposed to vast amounts of online text data during training.

This allows LLMs to display impressive language fluency, it also means they are capable of:

  1. Extrapolating information from the biases in training data,
  2. Misinterpreting ambiguous prompts, or modifying the information to align superficially with the input.

This becomes hugely alarming when language generation capabilities are used for sensitive applications, such as:

  1. Summarising medical records,
  2. Customer support conversations,
  3. Financial analysis reports, and providing erroneous legal advice.

 

Hallucination Mitigation Taxonomy

The study includes very insightful taxonomy of hallucination mitigation techniques for LLMs; both gradient and non-gradient.

Gradient approaches include complex and opaque decoding strategies, knowledge graphs, fine-tuning strategies and more.

Non-gradient approaches include RAG, Self-Refinement and prompt tuning.

Notably the RAG approaches are segmented into four parts;

  1. Before Generation
  2. During Generation
  3. After Generation
  4. End-to-End

The power of prompt engineering to mitigate hallucination lies in defining:

  1. Specific context &
  2. Expected Outcomes

 

The Best Defence

The best defence against hallucination is not one single approach or method, but a combined approach defending against hallucination.

Seamlessly integrating numerous mitigation approaches, is the most important takeaway.

 

The factors which any organisation should keep in mind are:

  1. To what extent is there a reliance on labeled data?

  2. What are the possibilities of introducing unsupervised or weak-supervised learning techniques to improve scalability and flexibility?

  3. Consideration of gradient and non-gradient approaches to produce coherent and contextually relevant information.

  4. The collected works on hallucination mitigation reveal a diverse array of strategies, each contributing uniquely to address the nuances of hallucination in LLMs.

  5. Self-refinement through feedback and reasoning brings forth impactful strategies.

  6. Structured Comparative reasoning introduces a structured approach to text preference prediction, enhancing coherence and reducing hallucination.

  7. Supervised fine-tuning can be explored via Knowledge Injection and Teacher-Student Approaches.

  8. Domain-specific knowledge is injected into weaker LLMs and approaches that employ counter factual datasets for improved factuality.

 

Find the study here: https://arxiv.org/pdf/2401.01313 

Previously published on Medium.

Subscribe to Blog Updates

START YOUR FREE TRIAL

Build powerful Virtual Assistants using Kore.ai Experience Optimization (XO) Platform.

 
Do you already have account? Login
By clicking Continue, 'you' agree to our Terms of Service
New call-to-action

Recent Posts

Follow us on LinkedIn

leftangle
Request a Demo Build a Virtual Assistant Resources