Raghav Singhal

I'm currently an AI PhD student in the Computer Science department at EPFL. I am fortunate to be advised by the amazing Prof. Robert West and am a contributor to the Apertus project (the biggest fully open and compliant training run & LLM to date).

Previously, I was a researcher at Massachusetts Institute of Technology and Mohamed bin Zayed University of Artificial Intelligence (MBZUAI), where I worked with Prof. Praneeth Vepakomma. Prior to this, I graduated from IIT Bombay with a Bachelor's in EE and a Master's in AI/ML.


For EPFL students: If you are interested in a project, please feel free to reach out via mail. I'm very happy to supervise motivated students!

Email  /  Scholar  /  Twitter  /  LinkedIn  /  Github

profile photo

Research

I spend my time these days working toward improving the foundations of foundation models, and studying fundamental flaws in current approaches to safety & alignment in LLMs.

Some of the projects I am currently working on include:

  • Instilling safety in LLMs during pretraining to create a strong prior for post-training, rather than relying on current post-training only approaches that are hacky and easy to break.
  • Improving the faithfulness of chain-of-thought reasoning in LLMs.

Check out my Google Scholar for a complete list of publications. * denotes equal contribution.

Apertus Apertus: Democratizing Open and Compliant LLMs for Global Language Environments
Project Apertus
arXiv / Hugging Face / Pretrain Code / Pretrain Data / Posttrain Code / Posttrain Data / Evals

The biggest fully open and compliant training run & LLM to date.

8B and 70B fully pretrained open-data open-weights models, multilingual in >1000 languages. Performance equivalent or better than corresponding Llama 3 sizes.

PontTuset Safety Subspaces are Not Linearly Distinct: A Fine-Tuning Case Study
Kaustubh Ponkshe*, Shaan Shah*, Raghav Singhal*, Praneeth Vepakomma
Abridged at Lock-LLM @ NeurIPS 2025
Under review
code / arXiv

We show that safety alignment in LLMs is not confined to distinct subspaces (but rather, highly entangled with general abality directions), thus fundamentally challenging the foundation of subspace-based defenses.

PontTuset FedEx-LoRA: Exact Aggregation for Federated and Efficient Fine-Tuning of Foundation Models
Raghav Singhal*, Kaustubh Ponkshe*, Praneeth Vepakomma
ACL 2025 - Oral (Top 2.2% of submitted papers)
project page / code / arXiv

We achieve exact aggregation in distributed fine-tuning of LLMs, consistently improving over SOTA.

PontTuset ABBA: Highly Expressive Hadamard Product Adaptation for Large Language Models
Raghav Singhal*, Kaustubh Ponkshe*, Rohit Vartak*, Praneeth Vepakomma
Abridged at ES-FOMO @ ICML 2025 - Spotlight (Top 9.5% of accepted papers)
Under review
project page / code / arXiv

We introduce ABBA, a PEFT method that enhances expressivity by decoupling low-rank updates from pre-trained weights via a Hadamard product, consistently improving over SOTA methods.

PontTuset Initialization using Update Approximation is a Silver Bullet for Extremely Efficient Low-Rank Fine-Tuning
Kaustubh Ponkshe*, Raghav Singhal*, Eduard Gorbunov, Alexey Tumanov, Samuel Horvath, Praneeth Vepakomma
Abridged at SCOPE @ ICLR 2025
Under review
project page / code / arXiv

We provably achieve the best approximation of full fine-tuning in low-rank spaces solely through clever initialization, outperforming LoRA while using up to 90x fewer parameters.

PontTuset Fed-SB: A Silver Bullet for Extreme Communication Efficiency and Performance in (Private) Federated LoRA Fine-Tuning
Raghav Singhal*, Kaustubh Ponkshe*, Rohit Vartak, Lav Varshney, Praneeth Vepakomma
Abridged at ES-FOMO @ ICML 2025
Under review
project page / code / arXiv

We set a new Pareto frontier for distributed fine-tuning of LLMs, achieving SOTA performance, stronger privacy guarantees, and up to 230x lower communication costs.

PontTuset M3CoL: Harnessing Shared Relations via Multimodal Mixup Contrastive Learning for Multimodal Classification
Raja Kumar*, Raghav Singhal*, Pranamya Kulkarni, Deval Mehta, Kshitij Jadhav
TMLR
project page / code / arXiv

We introduce a multimodal mixup-based contrastive learning framework that effectively captures shared relations across modalities, enabling robust multimodal representation learning.

PontTuset Regularization-based Framework for Quantization-, Fault- and Variability-Aware Training
Anmol Biswas*, Raghav Singhal*, Sivakumar Elangovan, Shreyas Sabnis, Udayan Ganguly
Abridged at MLNCP @ NeurIPS 2024
Under review
arXiv

We develop a learnable, non-uniform quantization-aware training framework that boosts efficiency and reliability of AI models deployed on low-power edge devices.

PontTuset Translation and Scale Invariance for Event-Based Object Tracking
Jens Egholm Pedersen, Raghav Singhal, Jörg Conradt
NICE 2023
code / paper

We train an extremely low-power SNN capable of accurate temporal regression, achieving ANN-level performance and faster convergence, directly portable to neuromorphic hardware.


Source code taken from Jon Barron's website.