No need for explanations: LLMs can implicitly learn from mistakes in-context
File(s)2502.08550v3.pdf (1.83 MB)
Accepted version
Author(s)
Type
Conference Paper
Abstract
Showing incorrect answers to Large Language Models (LLMs) is a popular strategy to improve their performance in reasoning-intensive tasks. It is widely assumed that, in order to be helpful, the incorrect answers must be accompanied by comprehensive rationales, explicitly detailing where the mistakes are and how to correct them. However, in this work we present a counterintuitive finding: we observe that LLMs perform *better* in math reasoning tasks when these rationales are eliminated from the context and models are left to infer on their own what makes an incorrect answer flawed. This approach also substantially outperforms chain-of-thought prompting in our evaluations. These results are consistent across LLMs of different sizes and varying reasoning abilities. To gain an understanding of *why* LLMs learn from mistakes more effectively without explicit corrective rationales, we perform a thorough analysis, investigating changes in context length and answer diversity between different prompting strategies, and their effect on performance. We also examine evidence of overfitting to the in-context rationales when these are provided, and study the extent to which LLMs are able to autonomously infer high-quality corrective rationales given only incorrect answers as input. We find evidence that, while incorrect answers are more beneficial for LLM learning than additional diverse *correct* answers, explicit corrective rationales over-constrain the model, thus limiting those benefits.
Date Issued
2025-11-04
Date Acceptance
2025-08-20
Citation
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP 2025), 2025, pp.33179-33203
Publisher
Association for Computational Linguistics
Start Page
33179
End Page
33203
Journal / Book Title
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP 2025)
Copyright Statement
© 2025 Association for Computational Linguistics
Source
EMNLP 2025
Publication Status
Published
Start Date
2025-11-04
Finish Date
2025-11-09
Coverage Spatial
Suzhou, China