Skip to content

jachansantiago/mode_guidance

Repository files navigation

MGD³: Mode-Guided Dataset Distillation using Diffusion Models

Project Page License

📌 ICML 2025 Oral (Top 1.0%)


🧠 Introduction

MGD³ presents a novel approach to dataset distillation by leveraging pre-trained diffusion models without the need for fine-tuning. The method enhances diversity and representativeness in synthetic datasets through a three-stage process:

  1. Mode Discovery: Identifies distinct data modes within each class.
  2. Mode Guidance: Steers the diffusion process toward the discovered modes.
  3. Stop Guidance: Transitions to unguided diffusion to prevent artifacts.

This approach ensures representative and diverse synthetic datasets suitable for training models.

For more details, visualizations, and supplementary materials, visit the Project Page.

🚀 Highlights

  • No Fine-Tuning Required: Utilizes pre-trained diffusion models directly.
  • Enhanced Diversity: Achieves superior intra-class diversity compared to existing methods.
  • Scalability: Demonstrates effectiveness on large-scale datasets like ImageNet-1K.

🛠️ Installation

  1. Clone the repository:
   git clone https://github.com/jachansantiago/mode_guidance.git
   cd mode_guidance
  1. Set up the environment:
   conda create -n modeguidance python=3.8
   conda activate modeguidance
   pip install -r requirements.txt
  1. For text-to-image distillation:

    Install our modified diffusers library:

   pip install -e diffusers

📊 Usage

To run the code on the ImageNette dataset:

./scripts/nette.sh

Acknowledgements

This project builds upon the following repositories:

About

Repository for Mode-Guided Latent Diffusion for Dataset Distillation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published