ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.08291
35
0

DreamFuse: Adaptive Image Fusion with Diffusion Transformer

11 April 2025
Junjia Huang
Pengxiang Yan
Jiyang Liu
Jie Wu
Zhao Wang
Yitong Wang
Liang Lin
G. Li
ArXivPDFHTML
Abstract

Image fusion seeks to seamlessly integrate foreground objects with background scenes, producing realistic and harmonious fused images. Unlike existing methods that directly insert objects into the background, adaptive and interactive fusion remains a challenging yet appealing task. It requires the foreground to adjust or interact with the background context, enabling more coherent integration. To address this, we propose an iterative human-in-the-loop data generation pipeline, which leverages limited initial data with diverse textual prompts to generate fusion datasets across various scenarios and interactions, including placement, holding, wearing, and style transfer. Building on this, we introduce DreamFuse, a novel approach based on the Diffusion Transformer (DiT) model, to generate consistent and harmonious fused images with both foreground and background information. DreamFuse employs a Positional Affine mechanism to inject the size and position of the foreground into the background, enabling effective foreground-background interaction through shared attention. Furthermore, we apply Localized Direct Preference Optimization guided by human feedback to refine DreamFuse, enhancing background consistency and foreground harmony. DreamFuse achieves harmonious fusion while generalizing to text-driven attribute editing of the fused results. Experimental results demonstrate that our method outperforms state-of-the-art approaches across multiple metrics.

View on arXiv
@article{huang2025_2504.08291,
  title={ DreamFuse: Adaptive Image Fusion with Diffusion Transformer },
  author={ Junjia Huang and Pengxiang Yan and Jiyang Liu and Jie Wu and Zhao Wang and Yitong Wang and Liang Lin and Guanbin Li },
  journal={arXiv preprint arXiv:2504.08291},
  year={ 2025 }
}
Comments on this paper