Maha Elbayad

AI Research Scientist at FAIR, Meta

profile_pic.jpg

I am a senior research scientist at Meta AI based in Menlo Park, CA. I specialize in massively multilingual and multimodal machine translation models for speech and text, working on projects like No-Language-Left-Behind and SeamlessM4T. I completed my PhD in 2020 from Université Grenoble Alpes, where I worked at Laboratoire d’informatique de Grenoble and Inria under the supervision of Jakob Verbeek and Laurent Besacier. My doctoral thesis explored novel designs of sequence-to-sequence models for efficient offline and streaming machine translation.

Prior to my PhD, I graduated from Centrale Paris (Applied Mathematics) and ENS Paris-Saclay (M2 MVA Mathématiques, Vision, Apprentisage).

News

Jan 15, 2025 SeamlessM4T published in Nature.
Dec 10, 2024 Publishing Large Concept Models (LCMs). This is a new direction in language modeling that moves beyond traditional token-level LLMs.
Nov 07, 2024 Merging Text Transformer Models from Different Initializations accepted by TMLR.
Jun 04, 2024 No Language Left Behind published in Nature.
Nov 30, 2023 Releasing Seamless communication models, a family of AI research models that enable more natural and authentic communication across languages (SeamlessExpressive, SeamlessStreaming and SeamlessM4T v2). Download the models from github or 🤗
Oct 24, 2023 TIME magazine selected SeamlessM4T as one of the best inventions of 2023!

Selected publications

  1. Joint speech and text machine translation for up to 100 languages
    Seamless Communication, Loı̈c Barrault, Yu-An Chung, Mariano Coria Meglioli, David Dale, and 64 more authors
    Nature, 2025
  2. Large Concept Models: Language Modeling in a Sentence Representation Space
    LCM Team, Loı̈c Barrault, Paul-Ambroise Duquenne, Maha Elbayad, Artyom Kozhevnikov, and 7 more authors
    arXiv e-prints, Dec 2024
  3. TMLR
    zipit.png
    Merging Text Transformer Models from Different Initializations
    Neha Verma, and Maha Elbayad
    Transactions on Machine Learning Research, Nov 2024
  4. Scaling neural machine translation to 200 languages
    Marta R. Costa-jussà, James Cross, Onur Çelebi, Maha Elbayad, Kenneth Heafield, and 34 more authors
    Nature, Jun 2024