Why reward models are still key to understanding alignment - podcast episode cover

Why reward models are still key to understanding alignment

Feb 14, 20248 minEp. 15
--:--
--:--
Listen in podcast apps:

Episode description

In an era dominated by direct preference optimization and LLMasajudge, why do we still need a model to output only a scalar reward?
This is AI generated audio with Python and 11Labs. Music generated by Meta's MusicGen.
Source code: https://github.com/natolambert/interconnects-tools
Original post: In an era dominated by direct preference optimization and LLM-as-a-judge, why do we still need a model to output only a scalar reward?

Podcast figures:
Figure 1: https://huggingface.co/datasets/natolambert/interconnects-figures/resolve/main/reward-models/img_004.png
Figure 2: https://huggingface.co/datasets/natolambert/interconnects-figures/resolve/main/reward-models/img_009.png

0:00 Why reward models are still key to understanding alignment



Get full access to Interconnects at www.interconnects.ai/subscribe
Why reward models are still key to understanding alignment | Interconnects podcast - Listen or read transcript on Metacast