Research Engineer - NLP & Large Language Models
Posted 5 hours 28 minutes ago by EPFL
Permanent
Not Specified
Research Jobs
Vaud, Lausanne, Switzerland
Job Description
EPFL, the Swiss Federal Institute of Technology in Lausanne, is one of theResearch Engineer - NLP & Large Language ModelsEPFL, the Swiss Federal Institute of Technology in Lausanne, is one of the most dynamic university campuses in Europe and ranks among the top 20 universities worldwide. The EPFL employs more than 6,500 people supporting the three main missions of the institution: research and The EPFL campus offers an exceptional working heart of a community of more than 17,000 people, including over 12,500 students and 4,000 researchers from X 120 different countries.Research Engineer - NLP & Large Language ModelsAbout the RoleWe are seeking a Research Engineer in Language Processing (NLP) and Large Language Models (LLMs) to contribute to the design, training, and of models. The role intersection of research and production-grade engineering, with a strong emphasis on post-training, multimodality, and advanced modeling techniques, including diffusion-based will work closely with researchers and applied scientists to novel ideas into scalable, reproducible systems, and to push the of the art in open, responsible, and multilingual ResponsibilitiesDesign, implement, and maintain training and post-training pipelines for large language and multimodal models ( , instruction tuning, alignment, preference research and engineering on post-training methodsContribute to multimodal modeling, text with modalities such as vision, speech, or audioExplore and apply diffusion-based models and hybrid approaches for language and multimodal learningOptimize large-scale training and inferenceDevelop pipelines and benchmarks for language understanding, reasoning, alignment, and multimodal with researchers to prototype new ideas, reproduce results from X contribute to or technical reportsEnsure code quality, reproducibility, and suitable for long-term X open-source releaseRequired or PhD in Computer Science, Machine Learning, AI, or a field (or equivalent practical experience)Strong background in NLP and deep learning, with hands-on experience working with large language modelsSolid programming skills in Python, with experience using modern ML frameworks ( , PyTorch)Experience working with open-weight or models, including releasing models, or benchmarksFamiliarity with post-training techniques for LLMs ( , instruction tuning, preference alignment)Strong experimental rigor: ability to design controlled experiments, analyze results, and efficientlyDesired / Bonus with diffusion models ( , text diffusion, diffusion, or multimodal diffusion)Hands-on work on multimodal models ( , text-image, text-audio, speech-language systems)Exposure to LLM alignment, safety, or beyond standard language modeling metricsExperience with distributed training and large-scale model with multilingual or low-resource language settingsContributions to open-source ML X research in NLP, multimodality, or We OfferA research-driven environment with access to large-scale compute and modern ML X leading researchers in NLP, multimodality, and modelingThe opportunity to work on open, and socially responsible AI systemsSupport for publishing research, contributing to open-source projects, and engaging with the broader research communityCompetitive and benefits, with Start : to be agreed uponActivity 1 year, renewableContract Type: Fixed-term contract (CDD) jid94f4a09aen jit0207aen jpiy26aen