Maria Lomeli
About me
I am a research engineer at Fundamental AI Research, Meta. Previously, I was a research scientist at Babylon Health, UK. Before that, I was a research associate, working with Zoubin Ghahramani at the Machine Learning group, CBL, University of Cambridge and member of Trinity Hall college.
I studied my PhD at the Gatsby Unit, UCL, my supervisor was Yee Whye Teh.
Journal publications
Petroni, F., Broscheit, S., Piktus, A., Lewis, P., Izacard, G., Hosseini, L., Dwivedi-Yu, J., Lomeli, M., Schick, T., Mazaré, P.E., Joulin, A., Grave, E., Riedel, S., ''Improving wikipedia verifiability with AI'' Nature Machine Intelligence, 2023, Vol 5, pp 1142–1148, NMI.
Izacard, G., Lewis, P., Lomeli, M., Hosseini, L., Petroni, F., Schick, T., Dwivedi-Yu, J., Joulin, A., Riedel, S., Grave, E., ''Atlas: few-shot learning with retrieval-augmented language models'' Journal of Machine Learning Research, 2023, Vol 24, pp 1-43, JMLR.
Mialon, G., Dessi, R., Lomeli, M., Nalmpantis, C., Pasunuru, R., Raileanu, R., Rosiere, B., Schick, T., Dwivedi-Yu, J., Celikyilmaz, A., LeCun, Y., Scialom, T., ''Augmented language models: a survey'', Transactions of Machine Learning Research, 2023, Vol 6, TLMR .
Valera, I., Pradier, M., Lomeli, M. and Ghahramani, Z., ''General Latent Feature Model for Heterogeneous Datasets'', Journal of Machine Learning Research, 2020, Vol 21 JMLR.
Lomeli, M., Rowland, M., Gretton, A. and Ghahramani, Z., ''Antithetic and Monte Carlo kernel estimators for partial rankings'', Statistics and Computing, 2019, Vol 29,1127–1147, StCo.
Lomeli, M., Favaro, S., Teh, Y. W.,'' A marginal sampler for
-Stable Poisson-Kingman mixture models'', Journal of Computational and Graphical Statistics, 2017, Vol 26,pp 44-53 JCGS.
Favaro, S., Lomeli, M., Nipoti, B., Teh, Y.W., ''Stick-breaking representations of
-stable Poisson-Kingman models'' , Electronic Journal of Statistics, 2014, Vol. 8, pp 1063-1085 EJS.
Favaro, S., Lomeli, M., Teh, Y.W.,''On a class of
-stable Poisson-Kingman models and an effective marginalized sampler'', Statistics and Computing, 2014, Vol 25, pp 67-78
StCo.
Proceedings
Cabannes, L., Beck, M., Szilvasy, G., Douze, M., Lomeli, M., Copet, J., Mazare, P., Synnaeve, G., and Jegou, H., 2026 ''Short window attention enables long-term memorization'' ICLR
Mekala, D., Weston, J., Lanchantin, J., Raileanu, R., Lomeli, M., Shang, J., Dwivedi-Yu, J., 2024, ''Toolverifier: Generalization to New Tools via Self-Verification '' EMNLP findings
Dwivedi-Yu, J., Schick, T., Jiang, Z., Lomeli, M., Lewis, P., Izacard, G., Grave, E., Riedel, S., Petroni, F., 2024, ''EditEval: an instruction-based benchmark for text improvements'' CoNLL
Lin, V. X., Chen, X., Chen, M., Shi, W., Lomeli, M., James, R., Rodriguez, P., Kahn, J., Szilvasy, G., Lewis, M., Zettlemoyer, L., Yih, S., 2024. ''RA-DIT: Retrieval-Augmented Dual Instruction Tuning'' ICLR
Shi, W., Min, S., Lomeli, M., Chou, Z., Li, M., Lin, V., Smith, N. A., Zettlemoyer, L., Yih, S., Lewis, M. 2024, ''In-Context Pretraining: Language Modeling Beyond Document Boundaries'' ICLR (accepted as spotlight)
Schick, T., Dwivedi-Yu, J., Dessi, R., Raileanu, R., Lomeli, M., Zettlemoyer, L., Cancedda, N., Scialom, T., 2023, ''Toolformer: language models can teach themselves to use tools'' Neural Information Processing Systems NeurIPS (accepted as an oral presentation)
Harman, M., Ahlgren, J., Berezin, M., Dulskyte, E., Dvortsova, I., George, J., Gucevska, N., Meijer, E., Spahr-Summers, J., Bojarczuk, K., Sapora, S. and Lomeli, M., 2021, ''Testing Web Enabled Simulation at Scale Using Metamorphic Testing'', International Conference on Software Engineering ICSE
Lomeli, M., Favaro, S.,Teh, Y.W., 2015, ''A hybrid sampler for Poisson-Kingman mixture models'', Neural information Processing Systems NeurIPS
Sejdinovic, D., Strathmann, H., Lomeli Garcia, M., Andrieu, C., Gretton, A., 2014,''Kernel Adaptive Metropolis-Hastings'', International Conference in Machine Learning ICML
Workshops
Lupidi, A., Gemmell, C., Cancedda, N., Dwivedi-Yu, J., Weston, J., Foerster, J., Raileanu,R. and Lomeli, M.,2024, ''Source2Synth: synthetic data generation and curation grounded in real data sources'' arXiv (best paper award SynthData workshop, ICLR2025)
Gautam, D., Lomeli, M., Gourgoulias, K., Thompson, D., Johri, S., 2019, ''Masking schemes for universal marginalisers'', Advances in Approximate Bayesian Inference symposium
Bloem-Reddy, B., Mathieu, E., Foster, A., Rainforth, T., Ge, H., Lomeli, M., Ghahramani, Z., Teh, Y.W., 2017, ''Sampling and inference for discrete random probability measures in probabilistic programs'', Advances in Approximate Bayesian Inference workshop, NeurIPS
Thesis
General Bayesian inference schemes in infinite mixture models
PhD thesis, University College London
UCL repository: Doctoral dissertation link
Preprints
Lomeli, M., Douze, M., Szilvasy, G., Cabannes, L., Copet, J., Sukhbaatar, S., Weston, J., Synnaeve, G., Mazare, P. and Jegou, H., 2025 ''Stochastic activations'' arXiv
Mazare, P., Szilvasy, G., Lomeli, M., Massa, F., Murray, N., Jegou, H., and Douze, M., 2025, ''Inference-time sparse attention with asymmetric indexing'' arXiv
Singh, A.K., Kocyigit, M. Y., Poulton, A., Esiobu, D., Lomeli, M., Szilvasy, G. and Hupkes, D., 2024, ''Evaluation data contamination in LLMs: how do we measure it and (when) does it matter?' arXiv
Douze, M., Guzhva, A., Deng, C., Johnson, J., Szilvasy, G., Mazare, P.E., Lomeli, M., Hosseini, L. and Jegou, H. 2024, ''The faiss library'' arXiv