ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2311.07069
  4. Cited By
Music ControlNet: Multiple Time-varying Controls for Music Generation

Music ControlNet: Multiple Time-varying Controls for Music Generation

13 November 2023
Shih-Lun Wu
Chris Donahue
Shinji Watanabe
Nicholas J. Bryan
    DiffMMGen
ArXiv (abs)PDFHTML

Papers citing "Music ControlNet: Multiple Time-varying Controls for Music Generation"

18 / 18 papers shown
Title
Video-Guided Text-to-Music Generation Using Public Domain Movie Collections
Video-Guided Text-to-Music Generation Using Public Domain Movie Collections
Haven Kim
Cheng-i Wang
Weihan Xu
Julian McAuley
Hao-Wen Dong
VGen
53
0
0
01 Jul 2025
Versatile Symbolic Music-for-Music Modeling via Function Alignment
Versatile Symbolic Music-for-Music Modeling via Function Alignment
Junyan Jiang
Daniel Y. Chin
Liwei Lin
Xuanjie Liu
Gus Xia
27
0
0
18 Jun 2025
LiLAC: A Lightweight Latent ControlNet for Musical Audio Generation
LiLAC: A Lightweight Latent ControlNet for Musical Audio Generation
Tom Baker
Javier Nistal
DiffM
28
0
0
13 Jun 2025
BNMusic: Blending Environmental Noises into Personalized Music
BNMusic: Blending Environmental Noises into Personalized Music
Chi Zuo
M. B. Møller
Pablo Martínez-Nuevo
Huayang Huang
Yu Wu
Ye Zhu
98
0
0
12 Jun 2025
Auto-Regressive vs Flow-Matching: a Comparative Study of Modeling Paradigms for Text-to-Music Generation
Or Tal
Felix Kreuk
Yossi Adi
AI4TS
60
0
0
10 Jun 2025
DRAGON: Distributional Rewards Optimize Diffusion Generative Models
DRAGON: Distributional Rewards Optimize Diffusion Generative Models
Yatong Bai
Jonah Casebeer
Somayeh Sojoudi
Nicholas J. Bryan
DiffMVLM
111
1
0
21 Apr 2025
TokenSynth: A Token-based Neural Synthesizer for Instrument Cloning and Text-to-Instrument
TokenSynth: A Token-based Neural Synthesizer for Instrument Cloning and Text-to-Instrument
Kyungsu Kim
Junghyun Koo
Sungho Lee
Haesun Joung
Kyogu Lee
139
0
0
13 Feb 2025
Editing Music with Melody and Text: Using ControlNet for Diffusion Transformer
Editing Music with Melody and Text: Using ControlNet for Diffusion Transformer
Siyuan Hou
Shansong Liu
Ruibin Yuan
Wei Xue
Ying Shan
Mangsuo Zhao
Chao Zhang
147
6
0
17 Jan 2025
SongEditor: Adapting Zero-Shot Song Generation Language Model as a Multi-Task Editor
SongEditor: Adapting Zero-Shot Song Generation Language Model as a Multi-Task Editor
Chenyu Yang
Shuai Wang
Hangting Chen
Jianwei Yu
Wei Tan
Rongzhi Gu
Yongjun Xu
Yizhi Zhou
Haina Zhu
Haoyang Li
KELM
422
2
0
18 Dec 2024
Generating Symbolic Music from Natural Language Prompts using an LLM-Enhanced Dataset
Generating Symbolic Music from Natural Language Prompts using an LLM-Enhanced Dataset
Weihan Xu
Julian McAuley
Taylor Berg-Kirkpatrick
Shlomo Dubnov
Hao-Wen Dong
121
1
0
02 Oct 2024
Video-Foley: Two-Stage Video-To-Sound Generation via Temporal Event Condition For Foley Sound
Video-Foley: Two-Stage Video-To-Sound Generation via Temporal Event Condition For Foley Sound
Junwon Lee
Jaekwon Im
Dabin Kim
Juhan Nam
VGen
138
10
0
21 Aug 2024
Combining audio control and style transfer using latent diffusion
Combining audio control and style transfer using latent diffusion
Andreas Maier
Yuliya Burankova
Anne Hartebrodt
David B. Blumenthal
DiffM
70
3
0
31 Jul 2024
Audio Conditioning for Music Generation via Discrete Bottleneck Features
Audio Conditioning for Music Generation via Discrete Bottleneck Features
Simon Rouard
Yossi Adi
Jade Copet
Axel Roebel
Alexandre Défossez
MGen
103
1
0
17 Jul 2024
Read, Watch and Scream! Sound Generation from Text and Video
Read, Watch and Scream! Sound Generation from Text and Video
Yujin Jeong
Yunji Kim
Sanghyuk Chun
Jiyoung Lee
VGenDiffM
89
15
0
08 Jul 2024
Subtractive Training for Music Stem Insertion using Latent Diffusion Models
Subtractive Training for Music Stem Insertion using Latent Diffusion Models
Ivan Villa-Renteria
Mason L. Wang
Zachary Shah
Zhe Li
Soohyun Kim
Neelesh Ramachandran
Mert Pilanci
181
0
0
27 Jun 2024
Diff-A-Riff: Musical Accompaniment Co-creation via Latent Diffusion
  Models
Diff-A-Riff: Musical Accompaniment Co-creation via Latent Diffusion Models
J. Nistal
Marco Pasini
Cyran Aouameur
M. Grachten
Stefan Lattner
DiffM
103
19
0
12 Jun 2024
DITTO-2: Distilled Diffusion Inference-Time T-Optimization for Music
  Generation
DITTO-2: Distilled Diffusion Inference-Time T-Optimization for Music Generation
Cheng-i Wang
Julian McAuley
Taylor Berg-Kirkpatrick
Nicholas J. Bryan
118
12
0
30 May 2024
SMITIN: Self-Monitored Inference-Time INtervention for Generative Music Transformers
SMITIN: Self-Monitored Inference-Time INtervention for Generative Music Transformers
Junghyun Koo
Gordon Wichern
François Germain
Sameer Khurana
Jonathan Le Roux
101
5
0
02 Apr 2024
1