Papers
arxiv:2510.14077

ERGO: Entropy-guided Resetting for Generation Optimization in Multi-turn Language Models

Published on Oct 15
· Submitted by Haziq Mohammad Khalid on Oct 20
Authors:
,
,
,
,
,

Abstract

ERGO, an entropy-guided resetting method, improves conversational AI performance by dynamically realigning context based on internal uncertainty, leading to enhanced accuracy and reliability in multi-turn interactions.

AI-generated summary

Large Language Models (LLMs) suffer significant performance degradation in multi-turn conversations when information is presented incrementally. Given that multi-turn conversations characterize everyday interactions with LLMs, this degradation poses a severe challenge to real world usability. We hypothesize that abrupt increases in model uncertainty signal misalignment in multi-turn LLM interactions, and we exploit this insight to dynamically realign conversational context. We introduce ERGO (Entropy-guided Resetting for Generation Optimization), which continuously quantifies internal uncertainty via Shannon entropy over next token distributions and triggers adaptive prompt consolidation when a sharp spike in entropy is detected. By treating uncertainty as a first class signal rather than a nuisance to eliminate, ERGO embraces variability in language and modeling, representing and responding to uncertainty. In multi-turn tasks with incrementally revealed instructions, ERGO yields a 56.6% average performance gain over standard baselines, increases aptitude (peak performance capability) by 24.7%, and decreases unreliability (variability in performance) by 35.3%, demonstrating that uncertainty aware interventions can improve both accuracy and reliability in conversational AI.

Community

Paper author Paper submitter

Multi-turn chats with LLMs often suffer from noise leading to major performance degradation when compared to optimal single-turn conversations. This paper introduces ERGO, which monitors token-level entropy and resets context when uncertainty spikes, essentially keeping the model “on track.” Across coding, SQL, math, and data-to-text tasks, ERGO boosts average performance by 56.6%, peak performance by 24.7% and reduces unreliability by 35.3%. Simple signal that generalizes across many tasks.

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2510.14077 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2510.14077 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2510.14077 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.