端到端的神经网络机器翻译(End-to-End Neural Machine Translation)是近几年兴起的一种全新的机器翻译方法。前面介绍了NMT的基本RNN Encoder-Decoder结构以及Attention Mechanism。但是经典的Encoder-Decoder结构存在一个明显的问题,就是源端与目标端都使用固定大小的词典,OOV词通常用一个UNK表示,如果目标端词典太大则会导致softmax计算复杂度变大。为了解决这一问题,研究者们也提出了一些工作。
转载请注明出处:http://blog.csdn.net/u011414416/article/details/51108193
本文将简要介绍如下工作:
Thang Luong, Ilya Sutskever, Quoc Le, Oriol Vinyals, and Wojciech Zaremba. Addressing the rare word problem in neural machine translation. In ACL 2015 (Long Paper).
Sébastien Jean, Kyunghyun Cho, Roland Memisevic, and Yoshua Bengio. On using very large target vocabulary for neural machine translation. In ACL 2015 (Long Paper).
Rohan Chitnis and John DeNero. Variable-length word encodings for neural translation models. in EMNLP 2015.
Wang Ling, Isabel Trancoso, Chris Dyer, and Alan Black. Charac