《Attention is All You Need》论文学习笔记
目录
Abstract
1. Illustrated Transformer
1.1 A High-level look
1.2 Attention
1.2.1 Scale Dot-Product Attention
1.2.2 Multi-Head Attention
1.3 Positional Encoding - Representing the Order of the Sequence
1.4 Detailed Architecture
References
最近在学习Pyt
原创
2020-07-07 10:43:37 ·
446 阅读 ·
0 评论