ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2104.08955
61
22
v1v2v3v4 (latest)

Many-Speakers Single Channel Speech Separation with Optimal Permutation Training

18 April 2021
Shaked Dovrat
Eliya Nachmani
Lior Wolf
    VLM
ArXiv (abs)PDFHTML
Abstract

Single channel speech separation has experienced great progress in the last few years. However, training neural speech separation for a large number of speakers (e.g., more than 10 speakers) is out of reach for the current methods, which rely on the Permutation Invariant Loss (PIT). In this work, we present a permutation invariant training that employs the Hungarian algorithm in order to train with an O(C3)O(C^3)O(C3) time complexity, where CCC is the number of speakers, in comparison to O(C!)O(C!)O(C!) of PIT based methods. Furthermore, we present a modified architecture that can handle the increased number of speakers. Our approach separates up to 202020 speakers and improves the previous results for large CCC by a wide margin.

View on arXiv
Comments on this paper