Risk-Averse Reinforcement Learning with Itakura-Saito Loss
Abstract
Proposed Itakura-Saito divergence-based loss function enhances numerical stability in risk-averse reinforcement learning using exponential utility functions.
Risk-averse reinforcement learning finds application in various high-stakes fields. Unlike classical reinforcement learning, which aims to maximize expected returns, risk-averse agents choose policies that minimize risk, occasionally sacrificing expected value. These preferences can be framed through utility theory. We focus on the specific case of the exponential utility function, where we can derive the Bellman equations and employ various reinforcement learning algorithms with few modifications. However, these methods suffer from numerical instability due to the need for exponent computation throughout the process. To address this, we introduce a numerically stable and mathematically sound loss function based on the Itakura-Saito divergence for learning state-value and action-value functions. We evaluate our proposed loss function against established alternatives, both theoretically and empirically. In the experimental section, we explore multiple financial scenarios, some with known analytical solutions, and show that our loss function outperforms the alternatives.
Community
The paper proposes a new loss function for risk-averse RL with exponential utility. The loss is mathematically justified (no euristics!) and numerically stable.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Risk-sensitive Reinforcement Learning Based on Convex Scoring Functions (2025)
- ShiQ: Bringing back Bellman to LLMs (2025)
- Moderate Actor-Critic Methods: Controlling Overestimation Bias via Expectile Loss (2025)
- Deep Reinforcement Learning Algorithms for Option Hedging (2025)
- Bellman operator convergence enhancements in reinforcement learning algorithms (2025)
- Taming OOD Actions for Offline Reinforcement Learning: An Advantage-Based Approach (2025)
- Imagination-Limited Q-Learning for Offline Reinforcement Learning (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper