ERGO: Entropy-guided Resetting for Generation Optimization in Multi-turn Language Models
Abstract
Large Language Models (LLMs) exhibit significant performance degradation in extended multi-turn interactions when task information is revealed incrementally, a common scenario in human-AI conversational settings. Such degradation presents critical challenges to maintaining consistency and reliability in real world multi-turn tasks. We hypothesize that abrupt increases in model uncertainty signal misalignment and impending conversational drift. To address this, we propose ERGO (Entropy-guided Resetting for Generation Optimization), an entropy‐guided framework that continuously monitors predictive entropy during multi‐turn exchanges and triggers adaptive prompt restructuring whenever entropy spikes. ERGO distills accumulated context into a concise, stateless prompt that preserves essential task details while discarding noise. Evaluated on diverse long‐horizon tasks, ERGO improves average multi‐turn performance by 56.6%, raises aptitude (peak performance) by 24.7%, and reduces unreliability (variability in performance) by 35.3%. By leveraging internal uncertainty as an alignment signal, ERGO offers a model‐agnostic, inference‐time intervention that enhances consistency, stability, and alignment in complex multi‐turn conversational AI systems.