Progress and Challenges in Addressing Hallucination Issues in Large Language Models

Authors

  • Zongke Li

DOI:

https://doi.org/10.54097/4a48mp77

Keywords:

Large Language Models, Hallucination Issues, Evaluation Methods, Mitigation Strategies.

Abstract

The illusion problem of large language models refers to the phenomenon where the content generated by the model is inconsistent with the input information or objective facts, significantly limiting its reliability and safety. This paper systematically reviews the definition and classification of the illusion problem, including factual illusions and fidelity illusions, intrinsic and extrinsic illusions, as well as types of closed-domain and open-domain illusions. It also summarizes the current mainstream evaluation methods from three perspectives: data, models, and multi-task applications. Regarding mitigation strategies, the article proposes various technical paths at the data, model, and application levels, including data cleaning and augmentation, model architecture optimization, prompt engineering, and real-time retrieval-augmented generation, which effectively enhance the accuracy and consistency of the generated content. Future research should aim to establish a more refined evaluation system, promote collaborative optimization across multiple technologies, and achieve dynamic knowledge updates and lightweight deployment to strengthen the practicality and safety of large models in real-world scenarios.

Downloads

Download data is not yet available.

References

[1] Ji ZW, et al. Survey of hallucination in natural language generation. ACM Computing Surveys, 2023, 55 (12): 248.

[2] Zhang S, Pan LM, Zhao JZ, Wang WY. The knowledge alignment problem: bridging human and external knowledge for large language models. arXiv: 2305.13669, 2024.

[3] Huang L, et al. A survey on hallucination in large language models: principles, taxonomy, challenges, and open questions. ACM Transactions on Information Systems, 2025, 43 (2).

[4] Liu ZY, Wang PJ, Song XB, Zhang X, Jiang BB. A survey on hallucination in large language models. Journal of Software, 2025, 36 (3): 1152 – 1185.

[5] Zhang Y, Li YF, Cui LY, Cai D, Liu LM, Fu TC, Huang XT, Zhao EB, Zhang Y, Chen YL, Wang LY, Luu AT, Bi W, Shi F, Shi SM. Siren’s song in the AI ocean: a survey on hallucination in large language models. arXiv: 2309.01219, 2023.

[6] Gardent C, Shimorina A, Narayan S, Perez-Beltrachini L. Creating training corpora for NLG micro-planning. In: Proc. of the 55th Annual Meeting of the Association for Computational Linguistics (Vol. 1: Long Papers). Vancouver: Association for Computational Linguistics, 2017: 179 – 188.

[7] Gabriel S, Celikyilmaz A, Jha R, Choi Y, Gao JF. GO FIGURE: a meta evaluation of factuality in summarization. In: Proceedings of the Association for Computational Linguistics: ACL - IJCNLP 2021. Association for Computational Linguistics, 2021: 478 – 487.

[8] Zhang W, Zhang J. Hallucination mitigation for retrieval-augmented large language models: a review. Mathematics, 2025, 13 (3): 856.

[9] Kumar K. Geotechnical Parrot Tales (GPT): harnessing large language models in geotechnical engineering. arXiv: 2304.02138, 2023.

[10] He J, Shen Y, Xie RF. Recognition and optimization of hallucination phenomena in large language models. Computer Applications, 2025, 45 (3): 709 – 714.

Downloads

Published

12-03-2026

How to Cite

Li, Z. (2026). Progress and Challenges in Addressing Hallucination Issues in Large Language Models. Highlights in Science, Engineering and Technology, 161, 121-127. https://doi.org/10.54097/4a48mp77