기본 정보
연구 분야
프로젝트
발행물
구성원
article|
gold
·인용수 0
·2026
Think Just Enough: Leveraging Self-Assessed Confidence for Adaptive Reasoning in Language Models
Junyeob Kim, Sang-goo Lee, Taeuk Kim
초록

Recent reinforcement learning (RL)-trained language models have demonstrated strong performance on complex reasoning tasks by producing long and detailed reasoning traces.However, despite these advancements, they often struggle with finding the right balance in reasoning length: some terminate prematurely before reaching a correct answer (underthinking), while others continue reasoning beyond necessity, leading to inefficiency or even degraded accuracy (overthinking).To address these challenges, we propose a method for optimizing reasoning length via self-assessed confidence.By prompting the model to evaluate its own confidence at intermediate reasoning steps, we enable dynamic stopping once sufficient reasoning is achieved.Experiments across multiple reasoning benchmarks show that our approach improves computational efficiency without compromising answer quality.Furthermore, we find that confidence estimates from RL-trained reasoning models are more reliable than those from standard LLMs, making it a valuable internal signal for controlling reasoning depth.

키워드
Language modelLanguage understandingNatural languageNon-monotonic logicAutomated reasoning
타입
article
IF / 인용수
- / 0
게재 연도
2026