ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.14404
10
0

Causally Steered Diffusion for Automated Video Counterfactual Generation

17 June 2025
Nikos Spyrou
Athanasios Vlontzos
Paraskevas Pegios
Thomas Melistas
Nefeli Gkouti
Yannis Panagakis
G. Papanastasiou
Sotirios A. Tsaftaris
    DiffMCMLVGen
ArXiv (abs)PDFHTML
Main:10 Pages
17 Figures
Bibliography:7 Pages
1 Tables
Appendix:8 Pages
Abstract

Adapting text-to-image (T2I) latent diffusion models for video editing has shown strong visual fidelity and controllability, but challenges remain in maintaining causal relationships in video content. Edits affecting causally dependent attributes risk generating unrealistic or misleading outcomes if these relationships are ignored. In this work, we propose a causally faithful framework for counterfactual video generation, guided by a vision-language model (VLM). Our method is agnostic to the underlying video editing system and does not require access to its internal mechanisms or finetuning. Instead, we guide the generation by optimizing text prompts based on an assumed causal graph, addressing the challenge of latent space control in LDMs. We evaluate our approach using standard video quality metrics and counterfactual-specific criteria, such as causal effectiveness and minimality. Our results demonstrate that causally faithful video counterfactuals can be effectively generated within the learned distribution of LDMs through prompt-based causal steering. With its compatibility with any black-box video editing system, our method holds significant potential for generating realistic "what-if" video scenarios in diverse areas such as healthcare and digital media.

View on arXiv
@article{spyrou2025_2506.14404,
  title={ Causally Steered Diffusion for Automated Video Counterfactual Generation },
  author={ Nikos Spyrou and Athanasios Vlontzos and Paraskevas Pegios and Thomas Melistas and Nefeli Gkouti and Yannis Panagakis and Giorgos Papanastasiou and Sotirios A. Tsaftaris },
  journal={arXiv preprint arXiv:2506.14404},
  year={ 2025 }
}
Comments on this paper