50
1

Reward-Guided Iterative Refinement in Diffusion Models at Test-Time with Applications to Protein and DNA Design

Abstract

To fully leverage the capabilities of diffusion models, we are often interested in optimizing downstream reward functions during inference. While numerous algorithms for reward-guided generation have been recently proposed due to their significance, current approaches predominantly focus on single-shot generation, transitioning from fully noised to denoised states. We propose a novel framework for inference-time reward optimization with diffusion models inspired by evolutionary algorithms. Our approach employs an iterative refinement process consisting of two steps in each iteration: noising and reward-guided denoising. This sequential refinement allows for the gradual correction of errors introduced during reward optimization. Besides, we provide a theoretical guarantee for our framework. Finally, we demonstrate its superior empirical performance in protein and cell-type-specific regulatory DNA design. The code is available at \href{this https URL}{this https URL}.

View on arXiv
@article{uehara2025_2502.14944,
  title={ Reward-Guided Iterative Refinement in Diffusion Models at Test-Time with Applications to Protein and DNA Design },
  author={ Masatoshi Uehara and Xingyu Su and Yulai Zhao and Xiner Li and Aviv Regev and Shuiwang Ji and Sergey Levine and Tommaso Biancalani },
  journal={arXiv preprint arXiv:2502.14944},
  year={ 2025 }
}
Comments on this paper