Alignment
-
RLHF
Reinforcement Learning from Human Feedback; uses human preference comparisons to fine-tune language models for safety and alignment.
-
Attention Mechanism
Weighted aggregation of context vectors, allowing models to focus on relevant information. Fundamental to transformers and modern NLP.