A Little Bit of Reinforcement Learning from Human Feedback
A short introduction to RLHF and post-training focused on language models.
Nathan Lambert
Chapter Contents
[Incomplete] Direct Alignment Algorithms
[Incomplete] Direct Alignment Algorithms
← Previous: Policy Gradients
Next: Constitutional AI →