ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.01639
48
1

SliderSpace: Decomposing the Visual Capabilities of Diffusion Models

3 February 2025
Rohit Gandikota
Zongze Wu
Richard Zhang
David Bau
Eli Shechtman
Nick Kolkin
    DiffM
ArXivPDFHTML
Abstract

We present SliderSpace, a framework for automatically decomposing the visual capabilities of diffusion models into controllable and human-understandable directions. Unlike existing control methods that require a user to specify attributes for each edit direction individually, SliderSpace discovers multiple interpretable and diverse directions simultaneously from a single text prompt. Each direction is trained as a low-rank adaptor, enabling compositional control and the discovery of surprising possibilities in the model's latent space. Through extensive experiments on state-of-the-art diffusion models, we demonstrate SliderSpace's effectiveness across three applications: concept decomposition, artistic style exploration, and diversity enhancement. Our quantitative evaluation shows that SliderSpace-discovered directions decompose the visual structure of model's knowledge effectively, offering insights into the latent capabilities encoded within diffusion models. User studies further validate that our method produces more diverse and useful variations compared to baselines. Our code, data and trained weights are available atthis https URL

View on arXiv
@article{gandikota2025_2502.01639,
  title={ SliderSpace: Decomposing the Visual Capabilities of Diffusion Models },
  author={ Rohit Gandikota and Zongze Wu and Richard Zhang and David Bau and Eli Shechtman and Nick Kolkin },
  journal={arXiv preprint arXiv:2502.01639},
  year={ 2025 }
}
Comments on this paper