Research
* denotes equal contribution
When Worse is Better: Navigating the Compression-Generation Tradeoff in Visual Tokenization
Vivek Ramanujan, Kushal Tirumala, Armen Aghajanyan, Luke Zettlemoyer, Ali Farhadi
NeurIPS 2025 — Spotlight
Studies when and how better image reconstruction leads to better generation. CRT achieves state-of-the-art ImageNet generation (2.18 FID) with 2-3x improved compute efficiency even with worse reconstruction.
From an Image to a Scene: Learning to Imagine the World from a Million 360 Videos
Matthew Wallingford, Anand Bhattad, Aditya Kusupati, Vivek Ramanujan, Matt Deitke, Sham Kakade, Aniruddha Kembhavi, Roozbeh Mottaghi, Wei-Chiu Ma, Ali Farhadi
NeurIPS 2024
360-1M dataset and Odin model for novel view synthesis from the largest real-world multi-view dataset to date.
The Unmet Promise of Synthetic Training Images: Using Retrieved Real Images Performs Better
Scott Geng, Cheng-Yu Hsieh, Vivek Ramanujan, Matthew Wallingford, Chun-Liang Li, Pang Wei Koh, Ranjay Krishna
NeurIPS 2024
Synthetic data can be beneficial, but is consistently matched or outperformed by real images from a simple retrieval baseline.
On the Connection between Pre-training Data Diversity and Fine-tuning Robustness
Vivek Ramanujan*, Thao Nguyen*, Sewoong Oh, Ludwig Schmidt, Ali Farhadi
NeurIPS 2023 — Spotlight
Data quantity is the primary factor influencing downstream robustness, while other factors have limited impact.
Neural Priming for Sample-Efficient Adaptation
Matthew Wallingford*, Vivek Ramanujan*, Alex Fang, Aditya Kusupati, Roozbeh Mottaghi, Aniruddha Kembhavi, Ludwig Schmidt, Ali Farhadi
NeurIPS 2023
Enabling large pretrained models to adapt to distribution shifts with minimal labeled data.
DataComp: In Search of the Next Generation of Multimodal Datasets
Samir Yitzhak Gadre*, Gabriel Ilharco*, Alex Fang*,... Jonathan Hayase, Georgios Smyrnis, Thao Nguyen, Ryan Marten, Mitchell Wortsman, Dhruba Ghosh, Jieyu Zhang, Eyal Orgad, Rahim Entezari, Giannis Daras, Sarah Pratt, Vivek Ramanujan,... Yonatan Bitton, Kalyani Marathe, Stephen Mussmann, Richard Vencu, Mehdi Cherti, Ranjay Krishna, Pang Wei Koh, Olga Saukh, Alexander Ratner, Shuran Song, Hannaneh Hajishirzi, Ali Farhadi, Romain Beaumont, Sewoong Oh, Alex Dimakis, Jenia Jitsev, Yair Carmon, Vaishaal Shankar, Ludwig Schmidt
NeurIPS 2023 — Datasets Track
A benchmark for multimodal dataset creation with 12.8B image-text pairs. DataComp-1B achieves 79.2% zero-shot ImageNet accuracy.
Neural Radiance Field Codebooks
Matthew Wallingford, Aditya Kusupati, Alex Fang, Vivek Ramanujan, Aniruddha Kembhavi, Roozbeh Mottaghi, Ali Farhadi
ICLR 2023
Learning object-centric representations through novel view reconstruction using a dictionary of object codes.
Matryoshka Representations for Adaptive Deployment
Aditya Kusupati, Gantavya Bhatt, Aniket Rege, Matthew Wallingford, Aditya Sinha, Vivek Ramanujan, William Howard-Snyder, Kaifeng Chen, Sham Kakade, Prateek Jain, Ali Farhadi
NeurIPS 2022
Flexible representations that adapt to multiple downstream tasks with varying computational resources.
Forward Compatible Training for Representation Learning
Vivek Ramanujan, Pavan Kumar Anasosalu Vasu, Ali Farhadi, Oncel Tuzel, Hadi Pouransari
CVPR 2022
Preparing for future model versions by saving cheap auxiliary information about present training.
Effects of Parameter Norm Growth During Transformer Training: Inductive Bias from Gradient Descent
Will Merrill, Vivek Ramanujan, Yoav Goldberg, Roy Schwartz, Noah Smith
EMNLP 2022 — Oral
As parameters grow in magnitude, networks approximate discretized networks with saturated activations—a new characterization of inductive bias in GD.
LLC: Accurate, Multi-Purpose Learnt Low-Dimensional Binary Codes
Aditya Kusupati, Matthew Wallingford, Vivek Ramanujan, Raghav Somani, Jae Sung Park, Krishna Pillutla, Prateek Jain, Sham Kakade, Ali Farhadi
NeurIPS 2021
Learning extremely low-dimensional binary codes (~20 bits for ImageNet-1K) while maintaining near-optimal accuracy.
Supermasks in Superposition
Mitchell Wortsman*, Vivek Ramanujan*, Rosanne Liu, Aniruddha Kembhavi, Mohammad Rastegari, Jason Yosinski, Ali Farhadi
NeurIPS 2020
Hidden networks for continual learning—learning thousands of tasks without catastrophic forgetting.
What's Hidden in a Randomly Weighted Neural Network?
Vivek Ramanujan*, Mitchell Wortsman*, Aniruddha Kembhavi, Ali Farhadi, Mohammad Rastegari
CVPR 2020
Finding untrained subnetworks at initialization that match trained network performance.
Soft Threshold Weight Reparameterization for Learnable Sparsity
Aditya Kusupati, Vivek Ramanujan, Raghav Somani, Mitchell Wortsman, Prateek Jain, Sham Kakade, Ali Farhadi
ICML 2020
A pruning strategy based on soft threshold reparametrization, enabling very sparse but highly performant trained models.
Improving Shape Deformation in Unsupervised Image-to-Image Translation
Aaron Gokaslan, Vivek Ramanujan, Kwang-In Kim, Daniel Ritchie, James Tompkin
ECCV 2018
Improving on CycleGAN by allowing better shape deformation between more disparate domains.