arXiv Analytics

Sign in

arXiv:2409.12799 [stat.ML]AbstractReferencesReviewsResources

The Central Role of the Loss Function in Reinforcement Learning

Kaiwen Wang, Nathan Kallus, Wen Sun

Published 2024-09-19Version 1

This paper illustrates the central role of loss functions in data-driven decision making, providing a comprehensive survey on their influence in cost-sensitive classification (CSC) and reinforcement learning (RL). We demonstrate how different regression loss functions affect the sample efficiency and adaptivity of value-based decision making algorithms. Across multiple settings, we prove that algorithms using the binary cross-entropy loss achieve first-order bounds scaling with the optimal policy's cost and are much more efficient than the commonly used squared loss. Moreover, we prove that distributional algorithms using the maximum likelihood loss achieve second-order bounds scaling with the policy variance and are even sharper than first-order bounds. This in particular proves the benefits of distributional RL. We hope that this paper serves as a guide analyzing decision making algorithms with varying loss functions, and can inspire the reader to seek out better loss functions to improve any decision making algorithm.

Related articles: Most relevant | Search more
arXiv:2205.07999 [stat.ML] (Published 2022-05-16)
An Exponentially Increasing Step-size for Parameter Estimation in Statistical Models
arXiv:2201.08536 [stat.ML] (Published 2022-01-21)
Instance-Dependent Confidence and Early Stopping for Reinforcement Learning
arXiv:2303.14281 [stat.ML] (Published 2023-03-24)
Sequential Knockoffs for Variable Selection in Reinforcement Learning