| Roberta: A Robustly Optimized BERT Pretraining Approach Y Liu, M Ott, N Goyal, J Du, M Joshi, D Chen, O Levy, M Lewis, ... arXiv preprint arXiv:1907.11692, 2019 | 24135 | 2019 |
| TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension M Joshi, E Choi, DS Weld, L Zettlemoyer Association for Computational Linguistics (ACL), 2017 | 3590 | 2017 |
| SpanBERT: Improving Pre-training by Representing and Predicting Spans M Joshi, D Chen, Y Liu, DS Weld, L Zettlemoyer, O Levy Transactions of the Association for Computational Linguistics, 2019 | 2650 | 2019 |
| Gemini 2.5: Pushing the frontier with advanced reasoning, multimodality, long context, and next generation agentic capabilities G Comanici, E Bieber, M Schaekermann, I Pasupat, N Sachdeva, I Dhillon, ... arXiv preprint arXiv:2507.06261, 2025 | 1374 | 2025 |
| BERT for Coreference Resolution: Baselines and Analysis M Joshi, O Levy, DS Weld, L Zettlemoyer Empirical Methods in Natural Language Processing (EMNLP), 2019 | 485 | 2019 |
| Roberta: A robustly optimized BERT pretraining approach, CoRR abs/1907.11692 (2019) Y Liu, M Ott, N Goyal, J Du, M Joshi, D Chen, O Levy, M Lewis, ... | 464* | 1907 |
| Pix2struct: Screenshot parsing as pretraining for visual language understanding K Lee, M Joshi, IR Turc, H Hu, F Liu, JM Eisenschlos, U Khandelwal, ... International Conference on Machine Learning, 18893-18912, 2023 | 435 | 2023 |
| Pali-x: On scaling up a multilingual vision and language model X Chen, J Djolonga, P Padlewski, B Mustafa, S Changpinyo, J Wu, ... arXiv preprint arXiv:2305.18565, 2023 | 277 | 2023 |
| Improving passage retrieval with zero-shot question generation D Sachan, M Lewis, M Joshi, A Aghajanyan, W Yih, J Pineau, ... Proceedings of the 2022 Conference on Empirical Methods in Natural Language …, 2022 | 225 | 2022 |
| Cm3: A causal masked multimodal model of the internet A Aghajanyan, B Huang, C Ross, V Karpukhin, H Xu, N Goyal, D Okhonko, ... arXiv preprint arXiv:2201.07520, 2022 | 188 | 2022 |
| DePlot: One-shot visual language reasoning by plot-to-table translation F Liu, J Eisenschlos, F Piccinno, S Krichene, C Pang, K Lee, M Joshi, ... Findings of the Association for Computational Linguistics: ACL 2023, 10381-10399, 2023 | 171 | 2023 |
| Matcha: Enhancing visual language pretraining with math reasoning and chart derendering F Liu, F Piccinno, S Krichene, C Pang, K Lee, M Joshi, Y Altun, N Collier, ... Proceedings of the 61st Annual Meeting of the Association for Computational …, 2023 | 149 | 2023 |
| An Information Bottleneck Approach for Controlling Conciseness in Rationale Extraction B Paranjape, M Joshi, J Thickstun, H Hajishirzi, L Zettlemoyer Empirical Methods in Natural Language Processing (EMNLP), 2020 | 138 | 2020 |
| From pixels to ui actions: Learning to follow instructions via graphical user interfaces P Shaw, M Joshi, J Cohan, J Berant, P Pasupat, H Hu, U Khandelwal, ... Advances in Neural Information Processing Systems 36, 34354-34370, 2023 | 110 | 2023 |
| Open-domain visual entity recognition: Towards recognizing millions of wikipedia entities H Hu, Y Luan, Y Chen, U Khandelwal, M Joshi, K Lee, K Toutanova, ... Proceedings of the IEEE/CVF International Conference on Computer Vision …, 2023 | 108 | 2023 |
| RoBERTa: A robustly optimized BERT pretraining approach (arXiv: 1907.11692). arXiv Y Liu, M Ott, N Goyal, J Du, M Joshi, D Chen, O Levy, M Lewis, ... | 105 | 1907 |
| A robustly optimized bert pretraining approach Y Liu, M Ott, N Goyal, J Du, M Joshi, D Chen, O Levy, M Lewis, ... arXiv preprint arXiv:1907.11692, 2019 | 90 | 2019 |
| Htlm: Hyper-text pre-training and prompting of language models A Aghajanyan, D Okhonko, M Lewis, M Joshi, H Xu, G Ghosh, ... arXiv preprint arXiv:2107.06955, 2021 | 87 | 2021 |
| triviaqa: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. arXiv e-prints, art M Joshi, E Choi, D Weld, L Zettlemoyer arXiv preprint arXiv:1705.03551, 2017 | 65 | 2017 |
| Cross-document coreference resolution over predicted mentions A Cattan, A Eirew, G Stanovsky, M Joshi, I Dagan arXiv preprint arXiv:2106.01210, 2021 | 63 | 2021 |