ERGO: Entropy-guided Resetting for Generation Optimization in Multi-turn Language Models
Abstract
ERGO, an entropy-guided resetting method, improves conversational AI performance by dynamically realigning context based on internal uncertainty, leading to enhanced accuracy and reliability in multi-turn interactions.
Large Language Models (LLMs) suffer significant performance degradation in multi-turn conversations when information is presented incrementally. Given that multi-turn conversations characterize everyday interactions with LLMs, this degradation poses a severe challenge to real world usability. We hypothesize that abrupt increases in model uncertainty signal misalignment in multi-turn LLM interactions, and we exploit this insight to dynamically realign conversational context. We introduce ERGO (Entropy-guided Resetting for Generation Optimization), which continuously quantifies internal uncertainty via Shannon entropy over next token distributions and triggers adaptive prompt consolidation when a sharp spike in entropy is detected. By treating uncertainty as a first class signal rather than a nuisance to eliminate, ERGO embraces variability in language and modeling, representing and responding to uncertainty. In multi-turn tasks with incrementally revealed instructions, ERGO yields a 56.6% average performance gain over standard baselines, increases aptitude (peak performance capability) by 24.7%, and decreases unreliability (variability in performance) by 35.3%, demonstrating that uncertainty aware interventions can improve both accuracy and reliability in conversational AI.
Community
Multi-turn chats with LLMs often suffer from noise leading to major performance degradation when compared to optimal single-turn conversations. This paper introduces ERGO, which monitors token-level entropy and resets context when uncertainty spikes, essentially keeping the model “on track.” Across coding, SQL, math, and data-to-text tasks, ERGO boosts average performance by 56.6%, peak performance by 24.7% and reduces unreliability by 35.3%. Simple signal that generalizes across many tasks.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Entropy-Guided Loop: Achieving Reasoning through Uncertainty-Aware Generation (2025)
- Drift No More? Context Equilibria in Multi-Turn LLM Interactions (2025)
- Think Just Enough: Sequence-Level Entropy as a Confidence Signal for LLM Reasoning (2025)
- Learning Contextual Retrieval for Robust Conversational Search (2025)
- SID: Multi-LLM Debate Driven by Self Signals (2025)
- ReSURE: Regularizing Supervision Unreliability for Multi-turn Dialogue Fine-tuning (2025)
- ARES: Multimodal Adaptive Reasoning via Difficulty-Aware Token-Level Entropy Shaping (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper