博客地址:http://blog.csdn.net/wangxinginnlp/article/details/56488921
准备在组内做一个关于神经机器翻译中语言学知识的引入(syntax + NMT)的报告,先把相关文章进行罗列下:
**下面都是个人观点
1. Linguistic Input Features Improve Neural Machine Translation (WMT2016)
http://www.statmt.org/wmt16/pdf/W16-2209.pdf
*提出用各种linguistic features来丰富encoding信息
*设计一种把linguistic features输入encoder的模式(虽然很简单)
2. Tree-to-Sequence Attentional Neural Machine Translation (ACL2016)
http://www.aclweb.org/anthology/P/P16/P16-1078.pdf
*设计一种tree-based encoder显示地把tree中间的节点表示出来,但tree-based encoder貌似只用到binary合成信息,没有用节点的类别信息(NP VP这种)
*attention mechanism在tree-based encoder和sequence encoder两种encoder的输出上分别做attention,然后把两种context vector简单相加形成最终的context vector供decoder使用
3. Multi-task sequenceto sequence learning (ICLR 2016)
https://arxiv.org/pdf/1511.06114.pdf
*One-to-many Setting中一个encoder,多个decoder分别用来做translation,parsing和auto-encode
*source-side syntax很巧妙的被利用:用source-side syntax做parsing,监督信号会被传到共享的encoder帮助更好地进行source sentence representation,从而帮助更好地进行translation,反之亦然