Uncertainty-Aware Reinforcement Learning for Robust Decision Making in LLM-Agent Collaboration

Authors

  • Maximilian R. Müller Department of Informatics, University of Heidelberg, 69120 Heidelberg, Germany Author
  • Tobias F. Weber Department of Informatics, University of Heidelberg, 69120 Heidelberg, Germany Author
  • Anna K. Schneider Department of Informatics, University of Heidelberg, 69120 Heidelberg, Germany Author

DOI:

https://doi.org/10.71465/fair776

Keywords:

Uncertainty modeling, Reinforcement learning, LLM agents, Robust decision-making, Bayesian optimization

Abstract

Decision-making processes involving LLM agents are often susceptible to uncertainty arising from ambiguous inputs and stochastic generation behavior. To address this issue, this study proposes an uncertainty-aware reinforcement learning framework that incorporates Bayesian reward estimation and entropy-regularized policy updates. The approach is validated on a benchmark consisting of 9,600 tasks with varying levels of input ambiguity, including incomplete instructions and conflicting objectives. Results indicate that the proposed method reduces error propagation by 28.7% and improves decision robustness, with task success rates increasing from 68.9% to 80.2%. Additionally, calibration metrics such as expected calibration error (ECE) decrease by 19.5%, demonstrating improved reliability in agent outputs. The framework provides a systematic solution for enhancing robustness in LLM-based collaborative systems under uncertainty.

Downloads

Download data is not yet available.

Downloads

Published

2026-04-01