Dissociation Of Faithful And Unfaithful Reasoning In Llms · The Large Language Model Bible Contribute to LLM-Bible

Dissociation Of Faithful And Unfaithful Reasoning In Llms

Yee Evelyn, Li Alice, Tang Chenyu, Jung Yeon Ho, Paturi Ramamohan, Bergen Leon. Arxiv 2024

[Paper]    
Interpretability And Explainability Uncategorized

Large language models (LLMs) often improve their performance in downstream tasks when they generate Chain of Thought reasoning text before producing an answer. We investigate how LLMs recover from errors in Chain of Thought. Through analysis of error recovery behaviors, we find evidence for unfaithfulness in Chain of Thought, which occurs when models arrive at the correct answer despite invalid reasoning text. We identify factors that shift LLM recovery behavior: LLMs recover more frequently from obvious errors and in contexts that provide more evidence for the correct answer. Critically, these factors have divergent effects on faithful and unfaithful recoveries. Our results indicate that there are distinct mechanisms driving faithful and unfaithful error recoveries. Selective targeting of these mechanisms may be able to drive down the rate of unfaithful reasoning and improve model interpretability.

Similar Work