[ NLP ] 自然语言处理必读论文及预训练模型(2019.10.28更)

【 Albert 】

ALBERT: A LITE BERT FOR SELF-SUPERVISED LEARNING OF LANGUAGE REPRESENTATIONS

论文链接:https://arxiv.org/pdf/1909.11942.pdf

GitHub:https://github.com/brightmart/albert_zh

 

【 SpanBERT 】

SpanBERT: Improving Pre-training by Representing and Predicting Spans

论文链接:https://arxiv.org/pdf/1907.10529.pdf

GitHub:https://github.com/facebookresearch/SpanBERT

 

【 RoBERTa 】

Liu Y, Ott M, Goyal N, et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach[J]. arXiv preprint arXiv:1907.11692, 2019.

论文链接:https://arxiv.org/pdf/1907.11692.pdf

GitHub: https://github.com/brightmart/roberta_zh

 

【 ERNIE 2.0 】

Ernie 2.0: A continual pre-training framework for language understanding

论文链接:https://arxiv.org/pdf/1907.12412.pdf?source=post_page

GitHub:https://github.com/PaddlePaddle/ERNIE

 

【 ERNIE 1.0 】

ERNIE: Enhanced Representation through Knowledge Integration

论文链接: https://arxiv.org/pdf/1904.09223.pdf

GitHub:https://github.com/PaddlePaddle/ERNIE

 

【 OpenAI GPT2 】

Radford, A., et al. (2019). "Language models are unsupervised multitask learners." OpenAI Blog 1(8).

{ GitHub: https://github.com/openai/gpt-2 }

 

【 XLNet 】

Yang, Z., et al. (2019). "XLNet: Generalized Autoregressive Pretraining for Language Understanding." arXiv preprint arXiv:1906.08237.

{ GitHub: https://github.com/zihangdai/xlnet }

{ Chinese-pretrained-model GitHub: https://github.com/ymcui/Chinese-PreTrained-XLNet }

 

【 BERT wwm - ext 】

{ GitHub: https://github.com/ymcui/Chinese-BERT-wwm }

 

【 BERT wwm 】

Cui, Y., et al. (2019). "Pre-Training with Whole Word Masking for Chinese BERT." arXiv preprint arXiv:1906.08101.

哈工大讯飞联合发布全词覆盖中文BERT预训练模型

{ GitHub: https://github.com/ymcui/Chinese-BERT-wwm }

 

【 BERT 】

Devlin, J., et al. (2018). "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2018." arXiv preprint arXiv:1810.04805.

{ GitHub: BERT }

 

【 ELMo 】

Peters, M. E., et al. (2018). "Deep contextualized word representations." arXiv preprint arXiv:1802.05365. NAACL2018最佳论文

 

【 OpenAI GPT 】

Radford, A., et al. (2018). "Improving language understanding by generative pre-training." URL https://s3-us-west-2. amazonaws. com/openai-assets/research-covers/languageunsupervised/language understanding paper. pdf.

 

【 word2vec (Skip-gram model)】

Mikolov, T., et al. (2013). Distributed representations of words and phrases and their compositionality. Advances in neural information processing systems. NIPS

 

【 Attention 】

Vaswani, A., et al. (2017). Attention is all you need. Advances in neural information processing systems.

 

中文词向量:

腾讯800W中文词 song yan老师出品 https://ai.tencent.com/ailab/nlp/embedding.html

  • 5
    点赞
  • 30
    收藏
    觉得还不错? 一键收藏
  • 1
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论 1
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值