claim
Recent research in Reinforcement Learning (RL) for model alignment focuses on dissecting the mechanisms of how RL alters model behavior, comparing optimization landscapes of different algorithms, and understanding the risks of reward hacking.

Authors

Sources

Referenced by nodes (1)