Attention Is All You Need

Proposed the Transformer model, a novel architecture using self-attention to improve sequence transduction tasks.

September 7, 2023 · 2 min · Chengyu Zhang