《Attention is All You Need》论文学习笔记
目录Abstract1. Illustrated Transformer1.1 A High-level look1.2 Attention1.2.1 Scale Dot-Product Attention1.2.2 Multi-Head Attention1.3 Positional Encoding - Representing the Order of the Sequence1.4 Detailed ArchitectureReferences最近在学习Pyt
原创
2020-07-07 10:43:37 ·
344 阅读 ·
0 评论