本文属于研究聊天机器人,前沿学术内容;附带大量 GitHub 程序,先收藏!
示例程序
Practical Neural Networks for NLP
- intro: EMNLP 2016
- github: https://github.com/clab/dynet_tutorial_examples
Structured Neural Networks for NLP: From Idea to Code
- slides: https://github.com/neubig/yrsnlp-2016/blob/master/neubig16yrsnlp.pdf
- github: https://github.com/neubig/yrsnlp-2016
Understanding Deep Learning Models in NLP
http://nlp.yvespeirsman.be/blog/understanding-deeplearning-models-nlp/
Deep learning for natural language processing, Part 1
https://softwaremill.com/deep-learning-for-nlp/
神经网络模型
Unifying Visual-Semantic Embeddings with Multimodal Neural Language Models
[外链图片转存失败,源站可能有防盗链机制,建议将图片保存下来直接上传(img-4VRLU1ok-1600087303353)(/assets/dl-materials/Unifying_Visual-Semantic_Embeddings_with_Multimodal_Neural_Language_Models.png)]
- arxiv: http://arxiv.org/abs/1411.2539
- github: https://github.com/ryankiros/visual-semantic-embedding
- results: http://www.cs.toronto.edu/~rkiros/lstm_scnlm.html
- demo: http://deeplearning.cs.toronto.edu/i2t
Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks
- arxiv: http://arxiv.org/abs/1503.00075
- github: https://github.com/stanfordnlp/treelstm
- github(Theano): https://github.com/ofirnachum/tree_rnn
Visualizing and Understanding Neural Models in NLP
- arxiv: http://arxiv.org/abs/1506.01066
- github: https://github.com/jiweil/Visualizing-and-Understanding-Neural-Models-in-NLP
Character-Aware Neural Language Models
Skip-Thought Vectors
A Primer on Neural Network Models for Natural Language Processing
Character-aware Neural Language Models
Neural Variational Inference for Text Processing
- arxiv: http://arxiv.org/abs/1511.06038
- notes: http://dustintran.com/blog/neural-variational-inference-for-text-processing/
- github: https://github.com/carpedm20/variational-text-tensorflow
- github: https://github.com/cheng6076/NVDM
Sequence to Sequence Learning
Generating Text with Deep Reinforcement Learning
- intro: NIPS 2015
- arxiv: http://arxiv.org/abs/1510.09202
MUSIO: A Deep Learning based Chatbot Getting Smarter
- homepage: http://ec2-204-236-149-143.us-west-1.compute.amazonaws.com:9000/
- github(Torch7): https://github.com/deepcoord/seq2seq
Translation
Learning phrase representations using rnn encoder-decoder for statistical machine translation
- intro: GRU. EMNLP 2014
- arxiv: http://arxiv.org/abs/1406.1078
Neural Machine Translation by Jointly Learning to Align and Translate
- intro: ICLR 2015
- arxiv: http://arxiv.org/abs/1409.0473
- github: https://github.com/lisa-groundhog/GroundHog
Multi-Source Neural Translation
- intro: “report up to +4.8 Bleu increases on top of a very strong attention-based neural translation model.”
- arxiv: [Multi-Source Neural Translation](Multi-Source Neural Translation)
- github(Zoph_RNN): https://github.com/isi-nlp/Zoph_RNN
- video: http://research.microsoft.com/apps/video/default.aspx?id=260336
Multi-Way, Multilingual Neural Machine Translation with a Shared Attention Mechanism
- arxiv: http://arxiv.org/abs/1601.01073
- github: https://github.com/nyu-dl/dl4mt-multi
- notes: https://github.com/dennybritz/deeplearning-papernotes/blob/master/notes/multi-way-nmt-shared-attention.md
Modeling Coverage for Neural Machine Translation
A Character-level Decoder without Explicit Segmentation for Neural Machine Translation
NEMATUS: Attention-based encoder-decoder model for neural machine translation
Variational Neural Machine Translation
- intro: EMNLP 2016
- arxiv: https://arxiv.org/abs/1605.07869
- github: https://github.com/DeepLearnXMU/VNMT
Neural Network Translation Models for Grammatical Error Correction
Linguistic Input Features Improve Neural Machine Translation
Sequence-Level Knowledge Distillation
- intro: EMNLP 2016
- arxiv: http://arxiv.org/abs/1606.07947
- github: https://github.com/harvardnlp/nmt-android
Neural Machine Translation: Breaking the Performance Plateau
Tips on Building Neural Machine Translation Systems
Semi-Supervised Learning for Neural Machine Translation
- intro: ACL 2016. Tsinghua University & Baidu Inc
- arxiv: http://arxiv.org/abs/1606.04596
EUREKA-MangoNMT: A C++ toolkit for neural machine translation for CPU
Deep Character-Level Neural Machine Translation
Neural Machine Translation Implementations
Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation
Learning to Translate in Real-time with Neural Machine Translation
Is Neural Machine Translation Ready for Deployment? A Case Study on 30 Translation Directions
Fully Character-Level Neural Machine Translation without Explicit Segmentation
Navigational Instruction Generation as Inverse Reinforcement Learning with Neural Machine Translation
Neural Machine Translation in Linear Time
- intro: ByteNet
- arxiv: https://arxiv.org/abs/1610.10099
- github: https://github.com/paarthneekhara/byteNet-tensorflow
- github(Tensorflow): https://github.com/buriburisuri/ByteNet
Neural Machine Translation with Reconstruction
A Convolutional Encoder Model for Neural Machine Translation
- intro: Facebook AI Research
- arxiv: https://arxiv.org/abs/1611.02344
Toward Multilingual Neural Machine Translation with Universal Encoder and Decoder
MXNMT: MXNet based Neural Machine Translation
Doubly-Attentive Decoder for Multi-modal Neural Machine Translation
- intro: Dublin City University & Trinity College Dublin
- arxiv: https://arxiv.org/abs/1702.01287
Massive Exploration of Neural Machine Translation Architectures
- intro: Google Brain
- arxiv: https://arxiv.org/abs/1703.03906
- github: https://github.com/google/seq2seq/
Depthwise Separable Convolutions for Neural Machine Translation
- intro: Google Brain & University of Toronto
- arxiv: https://arxiv.org/abs/1706.03059
自动摘要
Extraction of Salient Sentences from Labelled Documents
- arxiv: http://arxiv.org/abs/1412.6815
- github: https://github.com/mdenil/txtnets
- notes: https://github.com/jxieeducation/DIY-Data-Science/blob/master/papernotes/2014/06/model-visualizing-summarising-conv-net.md
A Neural Attention Model for Abstractive Sentence Summarization
- intro: EMNLP 2015. Facebook AI Research
- arxiv: http://arxiv.org/abs/1509.00685
- github: https://github.com/facebook/NAMAS
- github(TensorFlow): https://github.com/carpedm20/neural-summary-tensorflow
A Convolutional Attention Network for Extreme Summarization of Source Code
- homepage: http://groups.inf.ed.ac.uk/cup/codeattention/
- arxiv: http://arxiv.org/abs