Hallucinations in Scholarly LLMs
A Conceptual Overview and Practical Implications
DOI:
https://doi.org/10.52825/ocp.v8i.3175Keywords:
Hallucination, Large Language Models, Knowledge Graphs, Scholarly Communication, Neurosymbolic AI, Retrieval-Augmented GenerationAbstract
The issue of large language models (LLMs) is gradually infiltrating the academic workflow, but it also presents one significant problem: hallucination. The hallucinations involve invented research results, ideas of fabricated reference, and misinterpreted inferences that destroy the credibility and dependability of scholarly writing. In the present paper, the concept of hallucinations as the aspect of scholarly communication is discussed, the major types of hallucinations are revealed, and the causes along with effects of hallucinations are discussed. It also examines pragmatic mitigation measures, such as retrieval-augmented generation (RAG) of factual grounding, citation-verification, and neurosymbolic strategies of structured fact-checking. The paper additionally emphasizes the significance of human-AI partnership in the process of creating scholarly tools to make the use of AI in research responsible and verifiable. The paper seeks to create awareness and offer guidance to the creation of reliable AI systems to be used in scholarly contexts by synthesizing risks, opportunities, and available mitigation measures to such systems. Instead of presenting a comprehensive technical structure, the work provides an overview of the conceptual description which may be used to design more reliable, transparent, and fact-driven AI-assisted research tools.
Downloads
References
[1] J. G. Meyer, R. J. Urbanowicz, P. C. Martin et al., "ChatGPT and Large Language Models in Academia: Opportunities and Challenges", BioData Mining, vol. 16, no. 1, pp. 1–11, 2023. DOI: 10.1186/S13040-023-00339-9. [Online]. Available: https://link.springer.com/article/10.1186/s13040-023-00339-9.
[2] Y. Sun, D. Sheng, Z. Zhou, and Y. Wu, "AI Hallucination: Towards a Comprehensive Classification of Distorted Information in Artificial Intelligence-Generated Content", Humanities and Social Sciences Communications, vol. 11, no. 1, pp. 1–14, 2024. DOI: 10.1057/s41599-024-03811-x. [Online]. Available: https://www.nature.com/articles/s41599-024-03811-x.
[3] S. M. T. I. Tonmoy, S. M. M. Zaman, V. Jain et al., "A Comprehensive Survey of Hallucination Mitigation Techniques in Large Language Models", arXiv preprint arXiv:2401.01313, 2024. [Online]. Available: https://arxiv.org/pdf/2401.01313.
[4] Z. Bai, T. Xiao, Z. Han et al., "Hallucination of Multimodal Large Language Models: A Survey", arXiv preprint arXiv:2404.18930, 2024. [Online]. Available: https://arxiv.org/pdf/2404.18930.
[5] C. K. Bareh, "A Qualitative Assessment of the Accuracy of AI-LLM in Academic Research", AI and Ethics, vol. 5, no. 4, pp. 4305–4324, 2025. DOI: 10.1007/S43681-025-00730-8. [Online]. Available: https://link.springer.com/article/10.1007/s43681-025-00730-8.
[6] A. Jain, P. Nimonkar, and P. Jadhav, "Citation Integrity in the Age of AI: Evaluating the Risks of Reference Hallucination in Maxillofacial Literature", Journal of Cranio-Maxillofacial Surgery, vol. 53, no. 10, pp. 1871–1872, 2025. DOI: 10.1016/J.JCMS.2025.08.004.
[7] J. Niimi, "Hallucinations in Bibliographic Recommendation: Citation Frequency as a Proxy for Training Data Redundancy", arXiv preprint arXiv:2510.25378, 2025. [Online]. Available: https://arxiv.org/pdf/2510.25378.
[8] M. Chelli, J. Descamps, V. Lavoué et al., "Hallucination Rates and Reference Accuracy of ChatGPT and Bard for Systematic Reviews: Comparative Analysis", Journal of Medical Internet Research, vol. 26, no. 1, p. e53164, 2024. DOI: 10.2196/53164. [Online]. Available: https://www.jmir.org/2024/1/e53164.
[9] A. Cheng, V. Nagesh, S. Eller, V. Grant, and Y. Lin, "Exploring AI Hallucinations of ChatGPT: Reference Accuracy and Citation Relevance of ChatGPT Models and Training Conditions", Simulation in Healthcare, 2025. DOI: 10.1097/SIH.0000000000000877. [Online]. Available: https://journals.lww.com/simulationinhealthcare/fulltext/9900/exploring_ai_hallucinations_of_chatgpt__reference.198.aspx.
[10] O. Nsirim, "Hallucinations in Artificial Intelligence and Human Misinformation: Librarians' Perspectives on Implications for Scholarly Publication", Folia Toruniensia, vol. 25, pp. 79–98, 2025. DOI: 10.12775/FT.2025.004. [Online]. Available: https://apcz.umk.pl/FT/article/view/60574.
[11] N. E. Mustaffa, K. E. Lai, C. N. Preece, and F. Y. Wong, "A Bibliometric Review of Large Language Model Hallucination", International Journal of Research and Innovation in Social Science, vol. IX, no. 9, pp. 5025–5037, 2025. DOI: 10.47772/IJRISS.2025.909000409. [Online]. Available: https://rsisinternational.org/journals/ijriss/articles/a-bibliometric-review-of-large-language-model-hallucination/.
[12] J. Jamaluddin, N. Abd Gaffar, and N. S. S. Din, "Hallucination: A Key Challenge to Artificial Intelligence-Generated Writing", Malaysian Family Physician, vol. 18, p. 68, 2023. DOI: 10.51866/LTE.527. [Online]. Available: https://pmc.ncbi.nlm.nih.gov/articles/PMC10726751/.
[13] M. R. King, "An Update on AI Hallucinations: Not as Bad as You Remember or as You’ve Been Told", Cellular and Molecular Bioengineering, pp. 1–6, 2025. DOI: 10.1007/S12195-025-00874-X. [Online]. Available: https://link.springer.com/article/10.1007/s12195-025-00874-x.
[14] S. Gupta, "Retrieval-Augmented Generation and Hallucination in Large Language Models: A Scholarly Overview", Scholars Journal of Engineering and Technology, vol. 13, no. 5, pp. 328–330, 2025. DOI: 10.36347/sjet.2025.v13i05.003.
[15] E. Gumaan, "Theoretical Foundations and Mitigation of Hallucination in Large Language Models", arXiv preprint arXiv:2507.22915, 2025. [Online]. Available: https://arxiv.org/pdf/2507.22915.
[16] A. Guleria, K. Krishan, V. Sharma, and T. Kanchan, "ChatGPT: Ethical Concerns and Challenges in Academics and Research", Journal of Infection in Developing Countries, vol. 17, no. 9, pp. 1292–1299, 2023. DOI: 10.3855/JIDC.18738.
[17] Z. Sun, "Large Language Models in Peer Review: Challenges and Opportunities", Scientometrics, pp. 1–44, 2025. DOI: 10.1007/S11192-025-05440-W. [Online]. Available: https://link.springer.com/article/10.1007/s11192-025-05440-w.
[18] D. Li, B. Palfi, C. K. Zhang et al., "Overview of the SciHal25 Shared Task on Hallucination Detection for Scientific Content", in Proceedings of the 2025 Shared Tasks on Scientific Document Processing, 2025, pp. 307–315. DOI: 10.18653/V1/2025.SDP-1.29. [Online]. Available: https://aclanthology.org/2025.sdp-1.29/.
[19] W. Zhang, and J. Zhang, "Hallucination Mitigation for Retrieval-Augmented Large Language Models: A Review", Mathematics, vol. 13, no. 5, p. 856, 2025. DOI: 10.3390/MATH13050856. [Online]. Available: https://www.mdpi.com/2227-7390/13/5/856.
[20] J. Song, X. Wang, J. Zhu et al., "RAG-HAT: A Hallucination-Aware Tuning Pipeline for LLM in Retrieval-Augmented Generation", in Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing (Industry Track), 2024, pp. 1548–1558. DOI: 10.18653/V1/2024.EMNLP-INDUSTRY.113. [Online]. Available: https://aclanthology.org/2024.emnlp-industry.113/.
[21] V. Magesh, F. Surani, M. Dahl, M. Suzgun, C. D. Manning, and D. E. Ho, "Hallucination-Free? Assessing the Reliability of Leading AI Legal Research Tools", Journal of Empirical Legal Studies, vol. 22, no. 2, pp. 216–242, 2025. DOI: 10.1111/JELS.12413.
[22] Y. Saxena, A. Padia, M. S. Chaudhary, K. Gunaratna, S. Parthasarathy, and M. Gaur, "Ranking-Free RAG: Replacing Re-ranking with Selection in RAG for Sensitive Domains", arXiv preprint arXiv:2505.16014, 2025. [Online]. Available: https://arxiv.org/abs/2505.16014.
[23] T. R. Besold, A. S. d'Avila Garcez, L. C. Lamb, and others, "Neurosymbolic AI for EGI: AAAI Tutorial Series", https://nesy-egi.github.io/, 2025.
Downloads
Published
How to Cite
Conference Proceedings Volume
Section
License
Copyright (c) 2025 Naveen Lamba, Sanju Tiwari, Manas Gaur

This work is licensed under a Creative Commons Attribution 4.0 International License.
Accepted 2025-11-15
Published 2025-12-18