Saumya Malik

Saumya Malik

Predoctoral Researcher at the Allen Institute for Artificial Intelligence

About Me

I am applying to PhD programs in the 2025-2026 cycle!

Hi! I'm Saumya, a predoctoral young investigator on the AllenNLP and Olmo team at the Allen Institute for Artificial Intelligence (Ai2), advised by Nathan Lambert. Before that, I got my undergraduate degree at Princeton University in 2024, majoring in Computer Science and minoring in Linguistics and Cognitive Science. I was very fortunate to work with Professor Danqi Chen.

I do research in NLP, and I'm broadly interested in improving our understanding of the interplay and sequencing of training stages, data decisions, and evaluation signals to build language models that are more capable but also more controllable and aligned for people.

Check out some of my work on open language modeling, reward models, and instruction following below or on my Google Scholar!

Publications and Preprints

2025

Olmo 3 Technical Report

Team Olmo (incl. Saumya Malik, Core Contributor)

2025

RewardBench 2: Advancing Reward Model Evaluation

Saumya Malik, Valentina Pyatkin, Sander Land, Jacob Morrison, Noah A. Smith, Hannaneh Hajishirzi, Nathan Lambert

Preprint Under Review, 2025

2025

Generalizing Verifiable Instruction Following

Valentina Pyatkin, Saumya Malik, Victoria Graf, Hamish Ivison, Shengyi Huang, Pradeep Dasigi, Nathan Lambert, Hannaneh Hajishirzi

NeurIPS Datasets & Benchmarks, 2025. Also one of 10 benchmarks adopted in the Artificial Analysis Intelligence Index!

2025

Tulu 3: Pushing Frontiers in Open Language Model Post-Training

Nathan Lambert, Jacob Morrison, Valentina Pyatkin, Shengyi Huang, Hamish Ivison, Faeze Brahman, Lester James V. Miranda, Alisa Liu, Nouha Dziri, Shane Lyu, Yuling Gu, Saumya Malik, Victoria Graf, Jena D. Hwang, Jiangjiang Yang, Ronan Le Bras, Oyvind Tafjord, Chris Wilhelm, Luca Soldaini, Noah A. Smith, Yizhong Wang, Pradeep Dasigi, Hannaneh Hajishirzi

COLM, 2025

2025

2 OLMo 2 Furious

Team Olmo (incl. Saumya Malik)

COLM, 2025

2024

QuRating: Selecting High-Quality Data for Training Language Models

Alexander Wettig, Aatmik Gupta, Saumya Malik, Danqi Chen

ICML, 2024 (Spotlight Paper - top 1.98% of submissions)

2024

Lost in the Logic: An Evaluation of Large Language Models' Reasoning Capabilities on LSAT Logic Games

Saumya Malik

Preprint, 2024