PRISM: Diversifying Dataset Distillation by Decoupling Architectural Priors

Nov 13, 2025·
Brian Bernhard Moser
,
Shalini Sarode
,
Federico Raue
,
Krzysztof Adamkiewicz
,
Arundhati Shanbhag
,
Joachim Folz
Tobias Christian Nauen Tobias Christian Nauen
,
Andreas Dengel
PDF
Abstract
Dataset distillation (DD) promises compact yet faithful synthetic data, but existing approaches often inherit the inductive bias of a single teacher model. As dataset size increases, this bias drives generation toward overly smooth, homogeneous samples, reducing intra-class diversity and limiting generalization. We present PRISM (PRIors from diverse Source Models), a framework that disentangles architectural priors during synthesis. PRISM decouples the logit-matching and regularization objectives, supervising them with different teacher architectures: a primary model for logits and a stochastic subset for batch-normalization (BN) alignment. On ImageNet-1K, PRISM consistently and reproducibly outperforms single-teacher methods (e.g., SRe2L) and recent multi-teacher variants (e.g., G-VBSM) at low- and mid-IPC regimes. The generated data also show significantly richer intra-class diversity, as reflected by a notable drop in cosine similarity between features.
Type
Publication
arXiv preprint (arXiv)
publications

For more information, see the paper pdf.

Citation

If you use this work, please cite our paper:

@misc{moser2025prism,
  title = {PRISM: Diversifying Dataset Distillation by Decoupling Architectural
           Priors},
  author = {Brian B. Moser and Shalini Sarode and Federico Raue and Stanislav
            Frolov and Krzysztof Adamkiewicz and Arundhati Shanbhag and Joachim
            Folz and Tobias C. Nauen and Andreas Dengel},
  year = {2025},
  eprint = {2511.09905},
  archivePrefix = {arXiv},
  primaryClass = {cs.LG},
  url = {https://arxiv.org/abs/2511.09905},
}
Tobias Christian Nauen
Authors
PhD Student
I’m a researcher of artificial intelligence at DFKI and RPTU Kaiserslautern-Landau. My research interests include efficient deep learning, transformer models, multimodal learning, and computer vision. In my PhD project, my focus lies on the development of efficient transformer models for vision, language, and multimodal tasks.