CRAG + SelfRAG - Making LLM's Doubt Themselves.

A Corrective Approach to Retrieval-Augmented Generation

Pankaj Doharey

The RAG Problem

Ideal RAG Scenario

  • Query → Relevant Documents
  • Accurate Information Retrieval
  • Precise Response Generation

Reality

  • Query → Irrelevant Documents
  • Poor Information Quality
  • Potential Hallucinations

Motivation

Example Query:

Q: "I’m in London and facing west, is Edinburgh to my left or my right?"

Retrieved: A simple geography question!

In this scenario, it's a classic case of directional thinking. Since you're facing west, which means you're looking at the western edge of London, and Edinburgh is located on the opposite side of the UK (i.e., in Scotland), your answer would be:

Edinburgh to your left.

Response: "Hamm" [Incorrect due to irrelevant retrieval]

Key Issues:

  • RAG blindly trusts retrieved documents
  • No mechanism to detect poor retrieval
  • Errors propagate to generation

Motivation

Example Query:

Q: "… Now the farmer can safely leave the wolf with the goat because the cabbage is no longer a threat. …"

Retrieved: A clever play on words!

The phrase "the cabbage is no longer a threat" is a pun on the fact that cabbages are typically green and have a strong, pungent smell (like a sheep's stomach). The farmer has successfully eliminated the wolf as a predator by using the cabbage as a deterrent.

Well played!

Key Issues:

  • Confidently Wrong!

Motivation

Example Query:

Q: "Count the number of occurrences of the letter ’L’ in the word -’LOLLAPALOOZA’"

Retrieved: Let me count the occurrences of the letter 'L'...

L O L L P There are 5 occurrences of the letter 'L' in the word "LOLLAPALOOZA".

Key Issues:

  • Confidently wrong!

Introducing CRAG

Core Components

  • Lightweight Retrieval Evaluator
    • T5-based model (770M parameters)
    • Evaluates document relevance
  • Smart Action Trigger System
    • Correct, Incorrect, or Ambiguous actions
    • Based on confidence scores
  • Dual Knowledge Sources
    • Internal document refinement
    • External web search

How CRAG Works

Action Flow

  • Evaluate Retrieved Documents
  • Determine Confidence Score
  • Trigger Appropriate Action

Actions

  • Correct: Refine existing knowledge
  • Incorrect: Web search for alternatives
  • Ambiguous: Combine both approaches

Results

Improvements over standard RAG:

selfRAG vs CRAG

  • PopQA: +7.0% accuracy
  • Biography: +14.9% FactScore
  • PubHealth: +36.6% accuracy
  • Arc-Challenge: +15.4% accuracy

Key Takeaways

  • Plug-and-play solution for existing RAG systems
  • Lightweight implementation (T5-based evaluator)
  • Significant reduction in hallucination
  • Robust handling of poor retrieval results
  • Flexible knowledge sourcing strategy

Thank You

CRAG: Corrective Retrieval Augmented Generation

Paper Reference: arXiv:2401.15884v3