Reinforcement-Learning
-
RLHF
Reinforcement Learning from Human Feedback; uses human preference comparisons to fine-tune language models for safety and alignment.
Reinforcement Learning from Human Feedback; uses human preference comparisons to fine-tune language models for safety and alignment.