Attention Is All You NeedRevolutionary paper introducing the Transformer architecture that became the foundation for modern language models.