Neural
-
Transformer
Attention-based neural architecture without recurrence; enables efficient parallel training and strong performance on language tasks. Published by Vaswani et al., 2017.
-
Attention Mechanism
Weighted aggregation of context vectors, allowing models to focus on relevant information. Fundamental to transformers and modern NLP.