arXiv Analytics

Sign in

arXiv:2406.14324 [cs.LG]AbstractReferencesReviewsResources

Revealing the Learning Process in Reinforcement Learning Agents Through Attention-Oriented Metrics

Charlotte Beylier, Simon M. Hofmann, Nico Scherf

Published 2024-06-20, updated 2025-02-05Version 2

The learning process of a reinforcement learning (RL) agent remains poorly understood beyond the mathematical formulation of its learning algorithm. To address this gap, we introduce attention-oriented metrics (ATOMs) to investigate the development of an RL agent's attention during training. In a controlled experiment, we tested ATOMs on three variations of a Pong game, each designed to teach the agent distinct behaviours, complemented by a behavioural assessment. ATOMs successfully delineate the attention patterns of an agent trained on each game variation, and that these differences in attention patterns translate into differences in the agent's behaviour. Through continuous monitoring of ATOMs during training, we observed that the agent's attention developed in phases, and that these phases were consistent across game variations. Overall, we believe that ATOM could help improve our understanding of the learning processes of RL agents and better understand the relationship between attention and learning.

Comments: Workshop on Scientific Methods for Understanding Deep Learning, NeurIPS 2024
Categories: cs.LG
Related articles: Most relevant | Search more
arXiv:2211.11869 [cs.LG] (Published 2022-11-21)
Examining Policy Entropy of Reinforcement Learning Agents for Personalization Tasks
arXiv:2007.16045 [cs.LG] (Published 2020-07-30)
Moody Learners -- Explaining Competitive Behaviour of Reinforcement Learning Agents
arXiv:1902.03657 [cs.LG] (Published 2019-02-10)
A Bandit Framework for Optimal Selection of Reinforcement Learning Agents