Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2410.06554
Cited By
The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models
9 October 2024
Yanjun Chen
Dawei Zhu
Yirong Sun
Xinghao Chen
Wei Zhang
Xiaoyu Shen
ALM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"The Accuracy Paradox in RLHF: When Better Reward Models Don't Yield Better Language Models"
Title
No papers