Papers
arxiv:2502.00511

Bridging Internal Probability and Self-Consistency for Effective and Efficient LLM Reasoning

Published on Feb 1
Authors:
,
,
,
,

Abstract

Recent advancements in large language models (LLMs) have demonstrated remarkable reasoning capabilities. However, single-shot inference often yields unreliable results for complex reasoning tasks, leading researchers to explore multiple reasoning paths through methods such as perplexity and <PRE_TAG>self-consistency</POST_TAG>. In this paper, we present the first theoretical error decomposition analysis of these techniques, breaking down their error into estimation error and model error. Our analysis reveals a fundamental trade-off: perplexity methods suffer from substantial model error due to the absence of a proper consistency function, while <PRE_TAG>self-consistency</POST_TAG> exhibits high estimation error due to a slow error convergence rate. To overcome these limitations, we propose Reasoning-Pruning <PRE_TAG>Perplexity Consistency (RPC)</POST_TAG>. This approach combines Perplexity Consistency, which seamlessly integrates LLM perplexity with <PRE_TAG>self-consistency</POST_TAG>, and Reasoning Pruning, which eliminates low-probability reasoning paths to effectively prevent the degeneration of estimation error reduction. Theoretical analysis demonstrates that RPC not only accelerates the convergence rate of estimation error to an exponential level but also holds strong potential for further reducing model error. Extensive empirical evaluations on seven benchmark datasets confirm that RPC can significantly improve reasoning performance, sample efficiency, and confidence reliability.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2502.00511 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2502.00511 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2502.00511 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.