ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1809.07454
  4. Cited By
Conv-TasNet: Surpassing Ideal Time-Frequency Magnitude Masking for
  Speech Separation

Conv-TasNet: Surpassing Ideal Time-Frequency Magnitude Masking for Speech Separation

20 September 2018
Yi Luo
N. Mesgarani
ArXivPDFHTML

Papers citing "Conv-TasNet: Surpassing Ideal Time-Frequency Magnitude Masking for Speech Separation"

50 / 753 papers shown
Title
Multi-Channel Masking with Learnable Filterbank for Sound Source
  Separation
Multi-Channel Masking with Learnable Filterbank for Sound Source Separation
Wang Dai
A. Politis
Tuomas Virtanen
28
0
0
14 Mar 2023
Towards Real-Time Single-Channel Speech Separation in Noisy and
  Reverberant Environments
Towards Real-Time Single-Channel Speech Separation in Noisy and Reverberant Environments
Julian Neri
Sebastian Braun
14
1
0
14 Mar 2023
Guided Speech Enhancement Network
Guided Speech Enhancement Network
Yang Yang
Shao-fu Shih
Hakan Erdogan
J. Lin
C. Lee
Yunpeng Li
George Sung
Matthias Grundmann
30
6
0
13 Mar 2023
Online Binaural Speech Separation of Moving Speakers With a Wavesplit
  Network
Online Binaural Speech Separation of Moving Speakers With a Wavesplit Network
Cong Han
N. Mesgarani
34
4
0
13 Mar 2023
A two-stage speaker extraction algorithm under adverse acoustic
  conditions using a single-microphone
A two-stage speaker extraction algorithm under adverse acoustic conditions using a single-microphone
Aviad Eisenberg
Sharon Gannot
Shlomo E. Chazan
21
2
0
13 Mar 2023
Improving the Intent Classification accuracy in Noisy Environment
Improving the Intent Classification accuracy in Noisy Environment
Mohamed Nabih Ali
A. Brutti
Daniele Falavigna
18
0
0
12 Mar 2023
On Neural Architectures for Deep Learning-based Source Separation of
  Co-Channel OFDM Signals
On Neural Architectures for Deep Learning-based Source Separation of Co-Channel OFDM Signals
Gary C. F. Lee
Amir Weiss
A. Lancho
Yury Polyanskiy
G. Wornell
AI4TS
17
6
0
11 Mar 2023
Multi-Dimensional and Multi-Scale Modeling for Speech Separation
  Optimized by Discriminative Learning
Multi-Dimensional and Multi-Scale Modeling for Speech Separation Optimized by Discriminative Learning
Zhaoxi Mu
Xinyu Yang
Wenjing Zhu
19
5
0
07 Mar 2023
A Multi-Stage Triple-Path Method for Speech Separation in Noisy and
  Reverberant Environments
A Multi-Stage Triple-Path Method for Speech Separation in Noisy and Reverberant Environments
Zhaoxi Mu
Xinyu Yang
Xiangyuan Yang
Wenjing Zhu
13
5
0
07 Mar 2023
Scaling strategies for on-device low-complexity source separation with
  Conv-Tasnet
Scaling strategies for on-device low-complexity source separation with Conv-Tasnet
Mohamed Nabih Ali
Francesco Paissan
Daniele Falavigna
A. Brutti
23
2
0
06 Mar 2023
Hybrid Y-Net Architecture for Singing Voice Separation
Hybrid Y-Net Architecture for Singing Voice Separation
Rashen Fernando
Pamudu Ranasinghe
Udula Ranasinghe
J. Wijayakulasooriya
Pantaleon Perera
11
2
0
05 Mar 2023
Spectrogram Inversion for Audio Source Separation via Consistency,
  Mixing, and Magnitude Constraints
Spectrogram Inversion for Audio Source Separation via Consistency, Mixing, and Magnitude Constraints
P. Magron
Tuomas Virtanen
24
0
0
03 Mar 2023
Defending against Adversarial Audio via Diffusion Model
Defending against Adversarial Audio via Diffusion Model
Shutong Wu
Jiong Wang
Ming-Yu Liu
Weili Nie
Chaowei Xiao
DiffM
37
25
0
02 Mar 2023
Extending DNN-based Multiplicative Masking to Deep Subband Filtering for
  Improved Dereverberation
Extending DNN-based Multiplicative Masking to Deep Subband Filtering for Improved Dereverberation
Jean-Marie Lemercier
Julian Tobergte
Timo Gerkmann
15
2
0
01 Mar 2023
Reducing the Prior Mismatch of Stochastic Differential Equations for
  Diffusion-based Speech Enhancement
Reducing the Prior Mismatch of Stochastic Differential Equations for Diffusion-based Speech Enhancement
Bunlong Lay
Simon Welker
Julius Richter
Timo Gerkmann
DiffM
10
24
0
28 Feb 2023
3D Neural Beamforming for Multi-channel Speech Separation Against
  Location Uncertainty
3D Neural Beamforming for Multi-channel Speech Separation Against Location Uncertainty
Rongzhi Gu
Shi-Xiong Zhang
Dong Yu
9
2
0
27 Feb 2023
DFSNet: A Steerable Neural Beamformer Invariant to Microphone Array
  Configuration for Real-Time, Low-Latency Speech Enhancement
DFSNet: A Steerable Neural Beamformer Invariant to Microphone Array Configuration for Real-Time, Low-Latency Speech Enhancement
A. Kovalyov
Kashyap Patel
Issa Panahi
28
3
0
26 Feb 2023
Metric-oriented Speech Enhancement using Diffusion Probabilistic Model
Metric-oriented Speech Enhancement using Diffusion Probabilistic Model
Chen Chen
Yuchen Hu
Weiwei Weng
Chng Eng Siong
DiffM
40
19
0
23 Feb 2023
Unsupervised Noise adaptation using Data Simulation
Unsupervised Noise adaptation using Data Simulation
Chen Chen
Yuchen Hu
Heqing Zou
Linhui Sun
Chng Eng Siong
33
13
0
23 Feb 2023
MossFormer: Pushing the Performance Limit of Monaural Speech Separation
  using Gated Single-Head Transformer with Convolution-Augmented Joint
  Self-Attentions
MossFormer: Pushing the Performance Limit of Monaural Speech Separation using Gated Single-Head Transformer with Convolution-Augmented Joint Self-Attentions
Shengkui Zhao
Bin Ma
33
52
0
23 Feb 2023
Unifying Speech Enhancement and Separation with Gradient Modulation for
  End-to-End Noise-Robust Speech Separation
Unifying Speech Enhancement and Separation with Gradient Modulation for End-to-End Noise-Robust Speech Separation
Yuchen Hu
Chen Chen
Heqing Zou
Xionghu Zhong
Chng Eng Siong
47
16
0
22 Feb 2023
DasFormer: Deep Alternating Spectrogram Transformer for
  Multi/Single-Channel Speech Separation
DasFormer: Deep Alternating Spectrogram Transformer for Multi/Single-Channel Speech Separation
Shuo Wang
Xiangyu Kong
Xiulian Peng
H. Movassagh
Vinod Prakash
Yan Lu
26
11
0
21 Feb 2023
A Sidecar Separator Can Convert a Single-Talker Speech Recognition
  System to a Multi-Talker One
A Sidecar Separator Can Convert a Single-Talker Speech Recognition System to a Multi-Talker One
Lingwei Meng
Jiawen Kang
Mingyu Cui
Yuejiao Wang
Xixin Wu
Helen M. Meng
15
17
0
20 Feb 2023
Speech Enhancement with Multi-granularity Vector Quantization
Speech Enhancement with Multi-granularity Vector Quantization
Xiaokang Zhao
Qiu-shi Zhu
Jie Zhang
23
0
0
16 Feb 2023
Local spectral attention for full-band speech enhancement
Local spectral attention for full-band speech enhancement
Zhongshu Hou
Qi Hu
Kai-Jyun Chen
Jing Lu
30
0
0
11 Feb 2023
Multi-Source Diffusion Models for Simultaneous Music Generation and
  Separation
Multi-Source Diffusion Models for Simultaneous Music Generation and Separation
Giorgio Mariani
Irene Tallini
Emilian Postolache
Michele Mancusi
Luca Cosmo
Emanuele Rodolà
DiffM
30
37
0
04 Feb 2023
Neural Target Speech Extraction: An Overview
Neural Target Speech Extraction: An Overview
Kateřina Žmolíková
Marc Delcroix
Tsubasa Ochiai
K. Kinoshita
JanHonza'' vCernocký
Dong Yu
23
84
0
31 Jan 2023
Separate And Diffuse: Using a Pretrained Diffusion Model for Improving
  Source Separation
Separate And Diffuse: Using a Pretrained Diffusion Model for Improving Source Separation
Shahar Lutati
Eliya Nachmani
Lior Wolf
DiffM
36
14
0
25 Jan 2023
On Batching Variable Size Inputs for Training End-to-End Speech
  Enhancement Systems
On Batching Variable Size Inputs for Training End-to-End Speech Enhancement Systems
Philippe Gonzalez
T. S. Alstrøm
Tobias May
24
9
0
25 Jan 2023
Latent Autoregressive Source Separation
Latent Autoregressive Source Separation
Emilian Postolache
Giorgio Mariani
Michele Mancusi
Andrea Santilli
Luca Cosmo
Emanuele Rodolà
BDL
DRL
12
8
0
09 Jan 2023
ReVISE: Self-Supervised Speech Resynthesis with Visual Input for
  Universal and Generalized Speech Enhancement
ReVISE: Self-Supervised Speech Resynthesis with Visual Input for Universal and Generalized Speech Enhancement
Wei-Ning Hsu
Tal Remez
Bowen Shi
Jacob Donley
Yossi Adi
DiffM
27
12
0
21 Dec 2022
An Audio-Visual Speech Separation Model Inspired by
  Cortico-Thalamo-Cortical Circuits
An Audio-Visual Speech Separation Model Inspired by Cortico-Thalamo-Cortical Circuits
Kai Li
Fenghua Xie
Hang Chen
K. Yuan
Xiaolin Hu
34
14
0
21 Dec 2022
Towards Unified All-Neural Beamforming for Time and Frequency Domain
  Speech Separation
Towards Unified All-Neural Beamforming for Time and Frequency Domain Speech Separation
Rongzhi Gu
Shi-Xiong Zhang
Yuexian Zou
Dong Yu
AI4TS
22
24
0
16 Dec 2022
DeFT-AN: Dense Frequency-Time Attentive Network for Multichannel Speech
  Enhancement
DeFT-AN: Dense Frequency-Time Attentive Network for Multichannel Speech Enhancement
Dongheon Lee
Jung-Woo Choi
27
25
0
15 Dec 2022
Tackling the Cocktail Fork Problem for Separation and Transcription of
  Real-World Soundtracks
Tackling the Cocktail Fork Problem for Separation and Transcription of Real-World Soundtracks
Darius Petermann
G. Wichern
Aswin Shanmugam Subramanian
Zhong-Qiu Wang
Jonathan Le Roux
27
10
0
14 Dec 2022
Multi-Scale Feature Fusion Transformer Network for End-to-End Single
  Channel Speech Separation
Multi-Scale Feature Fusion Transformer Network for End-to-End Single Channel Speech Separation
Yinhao Xu
Jian Zhou
L. Tao
H. Kwan
30
0
0
14 Dec 2022
GPU-accelerated Guided Source Separation for Meeting Transcription
GPU-accelerated Guided Source Separation for Meeting Transcription
Desh Raj
Daniel Povey
Sanjeev Khudanpur
23
34
0
10 Dec 2022
Hyperbolic Audio Source Separation
Hyperbolic Audio Source Separation
Darius Petermann
G. Wichern
Aswin Shanmugam Subramanian
Jonathan Le Roux
24
10
0
09 Dec 2022
Deep neural network techniques for monaural speech enhancement: state of
  the art analysis
Deep neural network techniques for monaural speech enhancement: state of the art analysis
P. Ochieng
28
21
0
01 Dec 2022
A General Unfolding Speech Enhancement Method Motivated by Taylor's
  Theorem
A General Unfolding Speech Enhancement Method Motivated by Taylor's Theorem
Andong Li
Guochen Yu
C. Zheng
Wenzhe Liu
Xiaodong Li
45
10
0
30 Nov 2022
Deep Neural Mel-Subband Beamformer for In-car Speech Separation
Deep Neural Mel-Subband Beamformer for In-car Speech Separation
Vinay Kothapally
Yong-mei Xu
Meng Yu
Shizhong Zhang
Dong Yu
30
10
0
22 Nov 2022
TF-GridNet: Integrating Full- and Sub-Band Modeling for Speech
  Separation
TF-GridNet: Integrating Full- and Sub-Band Modeling for Speech Separation
Zhongqiu Wang
Samuele Cornell
Shukjae Choi
Younglo Lee
Byeonghak Kim
Shinji Watanabe
32
119
0
22 Nov 2022
LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
LA-VocE: Low-SNR Audio-visual Speech Enhancement using Neural Vocoders
Rodrigo Mira
Buye Xu
Jacob Donley
Anurag Kumar
Stavros Petridis
V. Ithapu
M. Pantic
20
13
0
20 Nov 2022
A Two-Stage Deep Representation Learning-Based Speech Enhancement Method
  Using Variational Autoencoder and Adversarial Training
A Two-Stage Deep Representation Learning-Based Speech Enhancement Method Using Variational Autoencoder and Adversarial Training
Yang Xiang
Jesper Lisby Højvang
M. Rasmussen
M. G. Christensen
DRL
23
5
0
16 Nov 2022
Array Configuration-Agnostic Personalized Speech Enhancement using
  Long-Short-Term Spatial Coherence
Array Configuration-Agnostic Personalized Speech Enhancement using Long-Short-Term Spatial Coherence
Yicheng Hsu
Yonghan Lee
M. Bai
21
2
0
16 Nov 2022
Hybrid Transformers for Music Source Separation
Hybrid Transformers for Music Source Separation
Simon Rouard
Francisco Massa
Alexandre Défossez
16
128
0
15 Nov 2022
Reverberation as Supervision for Speech Separation
Reverberation as Supervision for Speech Separation
R. Aralikatti
Christoph Boeddeker
G. Wichern
Aswin Shanmugam Subramanian
Jonathan Le Roux
22
7
0
15 Nov 2022
An Investigation of the Combination of Rehearsal and Knowledge
  Distillation in Continual Learning for Spoken Language Understanding
An Investigation of the Combination of Rehearsal and Knowledge Distillation in Continual Learning for Spoken Language Understanding
Umberto Cappellazzo
Daniele Falavigna
A. Brutti
CLL
22
2
0
15 Nov 2022
The Potential of Neural Speech Synthesis-based Data Augmentation for
  Personalized Speech Enhancement
The Potential of Neural Speech Synthesis-based Data Augmentation for Personalized Speech Enhancement
Anastasia Kuznetsova
Aswin Sivaraman
Minje Kim
32
3
0
14 Nov 2022
MedleyVox: An Evaluation Dataset for Multiple Singing Voices Separation
MedleyVox: An Evaluation Dataset for Multiple Singing Voices Separation
Chang-Bin Jeon
Hyeongi Moon
Keunwoo Choi
Ben Sangbae Chon
Kyogu Lee
12
5
0
14 Nov 2022
Previous
123...567...141516
Next