arXiv:2305.07036 [cs.LG]AbstractReferencesReviewsResources
GFlowNets with Human Feedback
Yinchuan Li, Shuang Luo, Yunfeng Shao, Jianye Hao
Published 2023-05-11Version 1
We propose the GFlowNets with Human Feedback (GFlowHF) framework to improve the exploration ability when training AI models. For tasks where the reward is unknown, we fit the reward function through human evaluations on different trajectories. The goal of GFlowHF is to learn a policy that is strictly proportional to human ratings, instead of only focusing on human favorite ratings like RLHF. Experiments show that GFlowHF can achieve better exploration ability than RLHF.
Related articles: Most relevant | Search more
Reinforcement Learning from Human Feedback with Active Queries
arXiv:2107.01969 [cs.LG] (Published 2021-07-05)
The MineRL BASALT Competition on Learning from Human Feedback
Rohin Shah et al.
arXiv:2406.02764 [cs.LG] (Published 2024-06-04)
Adaptive Preference Scaling for Reinforcement Learning with Human Feedback