Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language Models

Siren's Song in the AI Ocean: A Survey on Hallucination in Large Language Models”

Authors

  • Yue Zhang Soochow University
  • Yafu Li Zhejiang University
  • Leyang Cui Westlake University
  • Deng Cai Tencent AI Lab
  • Lemao Liu Tencent AI Lab
  • Tingchen Fu Renmin University of China
  • Xinting Huang Tencent AI Lab
  • Enbo Zhao Tencent AI Lab
  • Yu Zhang Soochow University
  • Yulong Chen Cambridge University
  • Longyue Wang Tencent AI Lab
  • Anh Tuan Luu Nanyang Technological University
  • Wei Bi Tencent AI Lab
  • Freda Shi Toyota Technological Institute at Chicago
  • Shuming Shi Tencent AI Lab

Keywords:

hallucination, llms, factuality

Abstract

While large language models (LLMs) have demonstrated remarkable capabilities across a range of downstream tasks, a significant concern revolves around their propensity to exhibit hallucinations: LLMs occasionally generate content that diverges from the user input, contradicts previously generated context, or misaligns with established world knowledge. This phenomenon poses a substantial challenge to the reliability of LLMs in real-world scenarios. In this paper, we survey recent efforts on the detection, explanation, and mitigation of hallucination, with an emphasis on the unique challenges posed by LLMs. We present taxonomies of the LLM hallucination phenomena and evaluation benchmarks, analyze existing approaches aiming at mitigating LLM hallucination, and discuss potential directions for future research.

Published

2026-01-22