Opt: Open pre-trained transformer language models S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ... arXiv preprint arXiv:2205.01068, 2022 | 2529 | 2022 |
Multilingual denoising pre-training for neural machine translation Y Liu arXiv preprint arXiv:2001.08210, 2020 | 1870 | 2020 |
Xi Victoria Lin, Todor Mihaylov, Myle Ott, Sam Shleifer, Kurt Shuster, Daniel Simig, Punit Singh Koura, Anjali Sridhar, Tianlu Wang, and Luke Zettlemoyer. 2022 S Zhang, S Roller, N Goyal, M Artetxe, M Chen, S Chen, C Dewan, ... Opt: Open pretrained transformer language models 1, 2022 | 836 | 2022 |
Multilingual translation with extensible multilingual pretraining and finetuning Y Tang, C Tran, X Li, PJ Chen, N Goyal, V Chaudhary, J Gu, A Fan arXiv preprint arXiv:2008.00401, 2020 | 396 | 2020 |
Few-shot learning with multilingual language models XV Lin, T Mihaylov, M Artetxe, T Wang, S Chen, D Simig, M Ott, N Goyal, ... arXiv preprint arXiv:2112.10668, 2021 | 385* | 2021 |
TWC LOGD: A portal for linked open government data ecosystems L Ding, T Lebo, JS Erickson, D DiFranzo, GT Williams, X Li, J Michaelis, ... Journal of Web Semantics 9 (3), 325-333, 2011 | 264 | 2011 |
Self-rewarding language models W Yuan, RY Pang, K Cho, S Sukhbaatar, J Xu, J Weston arXiv preprint arXiv:2401.10020, 2024 | 252 | 2024 |
Flowseq: Non-autoregressive conditional sequence generation with generative flow X Ma, C Zhou, X Li, G Neubig, E Hovy arXiv preprint arXiv:1909.02480, 2019 | 213 | 2019 |
Self-alignment with instruction backtranslation X Li, P Yu, C Zhou, T Schick, O Levy, L Zettlemoyer, J Weston, M Lewis arXiv preprint arXiv:2308.06259, 2023 | 176 | 2023 |
A corpus for multilingual document classification in eight languages H Schwenk, X Li arXiv preprint arXiv:1805.09821, 2018 | 167 | 2018 |
Multilingual translation from denoising pre-training Y Tang, C Tran, X Li, PJ Chen, N Goyal, V Chaudhary, J Gu, A Fan Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 …, 2021 | 149 | 2021 |
On evaluation of adversarial perturbations for sequence-to-sequence models P Michel, X Li, G Neubig, JM Pino arXiv preprint arXiv:1903.06620, 2019 | 148 | 2019 |
Multilingual speech translation with efficient finetuning of pretrained models X Li, C Wang, Y Tang, C Tran, Y Tang, J Pino, A Baevski, A Conneau, ... arXiv preprint arXiv:2010.12829, 2020 | 143 | 2020 |
Data-gov wiki: Towards linking government data L Ding, D DiFranzo, A Graves, JR Michaelis, X Li, DL McGuinness, ... 2010 AAAI spring symposium series, 2010 | 127 | 2010 |
Lifting the curse of multilinguality by pre-training modular transformers J Pfeiffer, N Goyal, XV Lin, X Li, J Cross, S Riedel, M Artetxe arXiv preprint arXiv:2205.06266, 2022 | 112 | 2022 |
Efficient large scale language modeling with mixtures of experts M Artetxe, S Bhosale, N Goyal, T Mihaylov, M Ott, S Shleifer, XV Lin, J Du, ... arXiv preprint arXiv:2112.10684, 2021 | 104 | 2021 |
Jingfei Du, et al. 2021. Few-shot learning with multilingual language models XV Lin, T Mihaylov, M Artetxe, T Wang, S Chen, D Simig, M Ott, N Goyal, ... arXiv preprint arXiv:2112.10668, 35-40, 2021 | 98 | 2021 |
Chameleon: Mixed-modal early-fusion foundation models C Team arXiv preprint arXiv:2405.09818, 2024 | 94 | 2024 |
TWC data-gov corpus: incrementally generating linked government data from data. gov L Ding, D DiFranzo, A Graves, JR Michaelis, X Li, DL McGuinness, ... Proceedings of the 19th international conference on World Wide Web, 1383-1386, 2010 | 82 | 2010 |
Do language models have beliefs? methods for detecting, updating, and visualizing model beliefs P Hase, M Diab, A Celikyilmaz, X Li, Z Kozareva, V Stoyanov, M Bansal, ... arXiv preprint arXiv:2111.13654, 2021 | 76 | 2021 |