CARE: A Framework for Correcting Numerical Hallucinations in LLM-Generated Financial Texts
- Authors
- Kim, Jian; Jung, Woohwan
- Issue Date
- Jul-2025
- Publisher
- Institute of Electrical and Electronics Engineers Inc.
- Keywords
- financial texts; hallucinations; large language model; semi-supervised learning
- Citation
- Proceedings - 2025 IEEE Conference on Artificial Intelligence, CAI 2025, pp 69 - 74
- Pages
- 6
- Indexed
- SCOPUS
- Journal Title
- Proceedings - 2025 IEEE Conference on Artificial Intelligence, CAI 2025
- Start Page
- 69
- End Page
- 74
- URI
- https://scholarworks.bwise.kr/erica/handle/2021.sw.erica/126224
- DOI
- 10.1109/CAI64502.2025.00018
- Abstract
- Large language models (LLMs) are increasingly used for various tasks in the financial domain. However, when generating text from documents with substantial numerical data, LLMs frequently produce numerical hallucinations. Existing methods to address these hallucinations mitigated the issue by regenerating entire texts based on feedback. Despite efforts to mitigate this, LLMs remain vulnerable to inaccuracies in handling numerical data. In this paper, we propose a framework, CARE, that corrects numerical hallucinations by replacing incorrect numbers with values from the source documents. Our approach improves accuracy and reduces costs by enabling targeted corrections instead of full-text regeneration. Experimental results show that CARE outperforms prior feedback-based methods as well as zero-shot and Chain-of-Thought (CoT) prompting. © 2025 IEEE.
- Files in This Item
-
Go to Link
- Appears in
Collections - COLLEGE OF COMPUTING > DEPARTMENT OF ARTIFICIAL INTELLIGENCE > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.