Skip to content

LM pretraining for generation, reading list, resources, conference mappings.

Notifications You must be signed in to change notification settings

FranxYao/Language-Model-Pretraining-for-Text-Generation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 

Repository files navigation

Language-Model-Pretraining-for-Text-Generation

LM pretraining for generation, reading list, resources, conference mappings.


  • Deep contextualized word representations (ELMo). NAACL 18

    • I just like ELMo and AI2.
  • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, NAACL 19

  • BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension.

    • Current SOTA on summarization
  • Language Models are Unsupervised Multitask Learners (GPT2)

  • Improving Language Understanding by Generative Pre-Training (GPT)

  • Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (T5)

  • Unified Language Model Pre-training for Natural Language Understanding and Generation (UniLM), NeurIPS 19

    • Li Dong, Nan Yang, Wenhui Wang, Furu Wei, Xiaodong Liu, Yu Wang, Jianfeng Gao, Ming Zhou, Hsiao-Wuen Hon. MSR
  • Text Summarization with Pretrained Encoders. ACL 19

  • Unsupervised Pretraining for Sequence to Sequence Learning, EMNLP 17

    • Prajit Ramachandran and Peter J. Liu and Quoc V. Le, Google Brain
    • The LM pretraining idea, before BERT. Effective on MT and Summarization.
  • MASS: Masked Sequence to Sequence Pre-training for Language Generation, ICML 19

    • Kaitao Song, Xu Tan, Tao Qin, Jianfeng Lu, Tie-Yan Liu, NJU and MSRA
  • Pre-trained Language Model Representations for Language Generation, NAACL 19

    • Sergey Edunov, Alexei Baevski, Michael Auli, FAIR
    • LM pretraining fine tuned on different size of MT data. The smaller the more performance gain.
  • Semi-Supervised Sequence Modeling with Cross-View Training. (CVT) EMNLP 18

    • Kevin Clark, Minh-Thang Luong, Christopher D. Manning, Quoc V. Le
    • Semi-supervised learning with multi-task learning. MT is one application. Also improvements on other tasks
  • Few-shot NLG with Pre-trained Language Model

    • Zhiyu Chen, Harini Eavani, Yinyin Liu, and William Yang Wang
    • Extreme few-shot supervision: 50 training instances get 26 BLEU
  • Cross-lingual Language Model Pretraining, Arxiv 19

    • Guillaume Lample, Alexis Conneau, FAIR
    • Application to Cross-lingual classification, Unsupervised machine translation, Supervised machine translation (low-resource), Low-resource language model
  • BERT has a Mouth, and It Must Speak: BERT as a Markov Random Field Language Model, Arxiv 19

    • Alex Wang, Kyunghyun Cho, NYU
    • Directly sample from BERT, not as effective.
  • Constant-Time Machine Translation with Conditional Masked Language Models (link)

    • Marjan Ghazvininejad, Omer Levy, Yinhan Liu and Luke Zettlemoyer
    • Iterative decoding: generate with BERT, then polish with BERT, then continue polish

About

LM pretraining for generation, reading list, resources, conference mappings.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published