Hierarchical Learning for Generation with Long Source Sequences.
CoRR, 2021
Recipes for Building an Open-Domain Chatbot.
,
,
,
,
,
,
,
,
,
,
Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, 2021
Multilingual Denoising Pre-training for Neural Machine Translation.
Trans. Assoc. Comput. Linguistics, 2020
SpanBERT: Improving Pre-training by Representing and Predicting Spans.
Trans. Assoc. Comput. Linguistics, 2020
Recipes for building an open-domain chatbot.
,
,
,
,
,
,
,
,
,
,
,
CoRR, 2020
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension.
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020
RoBERTa: A Robustly Optimized BERT Pretraining Approach.
CoRR, 2019
Constant-Time Machine Translation with Conditional Masked Language Models.
CoRR, 2019
Mask-Predict: Parallel Decoding of Conditional Masked Language Models.
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019
Cloze-driven Pretraining of Self-attention Networks.
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, 2019