|国家预印本平台
首页|The Memorization Problem: Can We Trust LLMs' Economic Forecasts?

The Memorization Problem: Can We Trust LLMs' Economic Forecasts?

The Memorization Problem: Can We Trust LLMs' Economic Forecasts?

来源:Arxiv_logoArxiv
英文摘要

Large language models (LLMs) cannot be trusted for economic forecasts during periods covered by their training data. We provide the first systematic evaluation of LLMs' memorization of economic and financial data, including major economic indicators, news headlines, stock returns, and conference calls. Our findings show that LLMs can perfectly recall the exact numerical values of key economic variables from before their knowledge cutoff dates. This recall appears to be randomly distributed across different dates and data types. This selective perfect memory creates a fundamental issue -- when testing forecasting capabilities before their knowledge cutoff dates, we cannot distinguish whether LLMs are forecasting or simply accessing memorized data. Explicit instructions to respect historical data boundaries fail to prevent LLMs from achieving recall-level accuracy in forecasting tasks. Further, LLMs seem exceptional at reconstructing masked entities from minimal contextual clues, suggesting that masking provides inadequate protection against motivated reasoning. Our findings raise concerns about using LLMs to forecast historical data or backtest trading strategies, as their apparent predictive success may merely reflect memorization rather than genuine economic insight. Any application where future knowledge would change LLMs' outputs can be affected by memorization. In contrast, consistent with the absence of data contamination, LLMs cannot recall data after their knowledge cutoff date.

Alejandro Lopez-Lira、Yuehua Tang、Mingyin Zhu

经济学财政、金融

Alejandro Lopez-Lira,Yuehua Tang,Mingyin Zhu.The Memorization Problem: Can We Trust LLMs' Economic Forecasts?[EB/OL].(2025-04-20)[2025-06-07].https://arxiv.org/abs/2504.14765.点此复制

评论