Papers
arxiv:2508.19827

Analysing Chain of Thought Dynamics: Active Guidance or Unfaithful Post-hoc Rationalisation?

Published on Aug 27
· Submitted by XingweiT on Aug 28
Authors:
,

Abstract

Investigation into Chain-of-Thought dynamics and faithfulness across various models reveals inconsistencies in their reliance on CoT and its alignment with actual reasoning.

AI-generated summary

Recent work has demonstrated that Chain-of-Thought (CoT) often yields limited gains for soft-reasoning problems such as analytical and commonsense reasoning. CoT can also be unfaithful to a model's actual reasoning. We investigate the dynamics and faithfulness of CoT in soft-reasoning tasks across instruction-tuned, reasoning and reasoning-distilled models. Our findings reveal differences in how these models rely on CoT, and show that CoT influence and faithfulness are not always aligned.

Community

Paper author Paper submitter

Our paper investigates the faithfulness of CoT on the soft-reasoning tasks based on instruction-tuned, multi-step reasoning, and distilled reasoning models. We designed two experiments: 1) forcing an answer at intermediate reasoning steps to measure the gold answer confidence; 2) adding cues to mislead the models to measure the variance in the gold answer confidence. We found CoT often serves as a post-hoc justification for the instruction-tuned LLMs, but distilled reasoning LLMs rely heavily on CoT. Moreover, we found that unfaithful CoTs can still provide active guidance.

This work has been accepted by EMNLP 2025 (Main).

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2508.19827 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2508.19827 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2508.19827 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.