Reinforcement Learning through Human Feedback - EXPLAINED! | RLHF

preview_player
Показать описание
We talk about reinforcement learning through human feedback. ChatGPT among other applications makes use of this.

ABOUT ME

PLAYLISTS FROM MY CHANNEL

MATH COURSES (7 day free trial)

OTHER RELATED COURSES (7 day free trial)
Рекомендации по теме
Комментарии
Автор

At 6:58, you have an error: PPO is not used to build the reward model.

theartofwar
Автор

Great video! I have a few questions:

1) Why do we need to manually train the reward model with human feedback if the point is to evaluate responses of another pretrained model? Can't we just cut out the reward model altogether, rate the responses directly using human feedback to generate a loss value for each response, then backpropagate on that? Does it require less human input to train the reward model than to train the GPT model directly?

2) When backpropagating the loss, do you need to do recurrent backpropagation for a number of steps that is the same as the length of the token output?

3) Does the loss value apply equally to every token that is output? Seems like this would overly punish some words e.g. if the question starts with "why" it's likely the response is going to start with "because" regardless of what comes after. Does RLHF only work with sentence embeddings rather than word embeddings?

neetpride
Автор

Brilliant Bro 👌. Excellent explanation. I never understand RLHF reading so many books and notes. Your examples are GREAT & simple to understand 👌
I am new to your channel and subscribed.

RameshKumar-ngnf
Автор

Sir, please make a video on function approximation in RL

sangeethashowrya
Автор

what about the generation of rewards, will there be another model to check the relativity of the answer and the precision of the answer, cause we have a lot of data

manigoyal
Автор

(1) supervised fine-tuning (SFT), (2) reward model (RM) training, and (3) reinforcement learning via proximal policy optimization (PPO) on this reward model explain me

thangarajr-qwwy
Автор

Acts as a randomizing factor depending on whom you are getting feedback from

manigoyal
Автор

haha quiz time again:


0) when the person knows me well
1)D
2)B if proper human feedback
3)C

xabaki
Автор

Aren't we users are the humans in feedback loop for openai

manigoyal
Автор

looking like indian but accent like britisher, where u from bro ?

harshsahu
Автор

The video is informative and good. but stop saying quiz time in an annoying way

aswinselva