Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2311.07069
Cited By
Music ControlNet: Multiple Time-varying Controls for Music Generation
13 November 2023
Shih-Lun Wu
Chris Donahue
Shinji Watanabe
Nicholas J. Bryan
DiffM
MGen
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Music ControlNet: Multiple Time-varying Controls for Music Generation"
18 / 18 papers shown
Title
Video-Guided Text-to-Music Generation Using Public Domain Movie Collections
Haven Kim
Cheng-i Wang
Weihan Xu
Julian McAuley
Hao-Wen Dong
VGen
53
0
0
01 Jul 2025
Versatile Symbolic Music-for-Music Modeling via Function Alignment
Junyan Jiang
Daniel Y. Chin
Liwei Lin
Xuanjie Liu
Gus Xia
27
0
0
18 Jun 2025
LiLAC: A Lightweight Latent ControlNet for Musical Audio Generation
Tom Baker
Javier Nistal
DiffM
28
0
0
13 Jun 2025
BNMusic: Blending Environmental Noises into Personalized Music
Chi Zuo
M. B. Møller
Pablo Martínez-Nuevo
Huayang Huang
Yu Wu
Ye Zhu
98
0
0
12 Jun 2025
Auto-Regressive vs Flow-Matching: a Comparative Study of Modeling Paradigms for Text-to-Music Generation
Or Tal
Felix Kreuk
Yossi Adi
AI4TS
60
0
0
10 Jun 2025
DRAGON: Distributional Rewards Optimize Diffusion Generative Models
Yatong Bai
Jonah Casebeer
Somayeh Sojoudi
Nicholas J. Bryan
DiffM
VLM
111
1
0
21 Apr 2025
TokenSynth: A Token-based Neural Synthesizer for Instrument Cloning and Text-to-Instrument
Kyungsu Kim
Junghyun Koo
Sungho Lee
Haesun Joung
Kyogu Lee
139
0
0
13 Feb 2025
Editing Music with Melody and Text: Using ControlNet for Diffusion Transformer
Siyuan Hou
Shansong Liu
Ruibin Yuan
Wei Xue
Ying Shan
Mangsuo Zhao
Chao Zhang
147
6
0
17 Jan 2025
SongEditor: Adapting Zero-Shot Song Generation Language Model as a Multi-Task Editor
Chenyu Yang
Shuai Wang
Hangting Chen
Jianwei Yu
Wei Tan
Rongzhi Gu
Yongjun Xu
Yizhi Zhou
Haina Zhu
Haoyang Li
KELM
422
2
0
18 Dec 2024
Generating Symbolic Music from Natural Language Prompts using an LLM-Enhanced Dataset
Weihan Xu
Julian McAuley
Taylor Berg-Kirkpatrick
Shlomo Dubnov
Hao-Wen Dong
121
1
0
02 Oct 2024
Video-Foley: Two-Stage Video-To-Sound Generation via Temporal Event Condition For Foley Sound
Junwon Lee
Jaekwon Im
Dabin Kim
Juhan Nam
VGen
138
10
0
21 Aug 2024
Combining audio control and style transfer using latent diffusion
Andreas Maier
Yuliya Burankova
Anne Hartebrodt
David B. Blumenthal
DiffM
70
3
0
31 Jul 2024
Audio Conditioning for Music Generation via Discrete Bottleneck Features
Simon Rouard
Yossi Adi
Jade Copet
Axel Roebel
Alexandre Défossez
MGen
103
1
0
17 Jul 2024
Read, Watch and Scream! Sound Generation from Text and Video
Yujin Jeong
Yunji Kim
Sanghyuk Chun
Jiyoung Lee
VGen
DiffM
89
15
0
08 Jul 2024
Subtractive Training for Music Stem Insertion using Latent Diffusion Models
Ivan Villa-Renteria
Mason L. Wang
Zachary Shah
Zhe Li
Soohyun Kim
Neelesh Ramachandran
Mert Pilanci
181
0
0
27 Jun 2024
Diff-A-Riff: Musical Accompaniment Co-creation via Latent Diffusion Models
J. Nistal
Marco Pasini
Cyran Aouameur
M. Grachten
Stefan Lattner
DiffM
103
19
0
12 Jun 2024
DITTO-2: Distilled Diffusion Inference-Time T-Optimization for Music Generation
Cheng-i Wang
Julian McAuley
Taylor Berg-Kirkpatrick
Nicholas J. Bryan
118
12
0
30 May 2024
SMITIN: Self-Monitored Inference-Time INtervention for Generative Music Transformers
Junghyun Koo
Gordon Wichern
François Germain
Sameer Khurana
Jonathan Le Roux
101
5
0
02 Apr 2024
1