arXiv Analytics

Sign in

arXiv:2502.03752 [cs.LG]AbstractReferencesReviewsResources

PRISM: A Robust Framework for Skill-based Meta-Reinforcement Learning with Noisy Demonstrations

Sanghyeon Lee, Sangjun Bae, Yisak Park, Seungyul Han

Published 2025-02-06Version 1

Meta-reinforcement learning (Meta-RL) facilitates rapid adaptation to unseen tasks but faces challenges in long-horizon environments. Skill-based approaches tackle this by decomposing state-action sequences into reusable skills and employing hierarchical decision-making. However, these methods are highly susceptible to noisy offline demonstrations, resulting in unstable skill learning and degraded performance. To overcome this, we propose Prioritized Refinement for Skill-Based Meta-RL (PRISM), a robust framework that integrates exploration near noisy data to generate online trajectories and combines them with offline data. Through prioritization, PRISM extracts high-quality data to learn task-relevant skills effectively. By addressing the impact of noise, our method ensures stable skill learning and achieves superior performance in long-horizon tasks, even with noisy and sub-optimal data.

Comments: 8 pages main, 19 pages appendix with reference. Submitted to ICML 2025
Categories: cs.LG, cs.AI
Related articles: Most relevant | Search more
arXiv:2204.11828 [cs.LG] (Published 2022-04-25)
Skill-based Meta-Reinforcement Learning
arXiv:2006.07808 [cs.LG] (Published 2020-06-14)
Reinforcement Learning with Supervision from Noisy Demonstrations
arXiv:2306.05501 [cs.LG] (Published 2023-06-08)
AMEE: A Robust Framework for Explanation Evaluation in Time Series Classification