Announcing our new Paper: The Prompt Report, with Co-authors from OpenAI & Microsoft!

Check it out →

Definition of Reinforcement Learning from Human Feedback (RLHF)

RLHF is a method for fine tuning LLMs according to human preference data.

For more definitions, view the full list.

Word count: 0

Get AI Certified by Learn Prompting


Copyright © 2024 Learn Prompting.