ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.08410
31
4
v1v2v3v4v5 (latest)

Gaussian Mixture Reduction with Composite Transportation Divergence

19 February 2020
Qiong Zhang
Archer Gong Zhang
Jiahua Chen
ArXiv (abs)PDFHTML
Abstract

Gaussian mixtures can approximate almost any smooth density function and are used to simplify downstream inference tasks. As such, it is widely used in applications in density estimation, belief propagation, and Bayesian filtering. In these applications, a finite Gaussian mixture provides an initial approximation to density functions that are updated recursively. A challenge in these recursions is that the order of the Gaussian mixture increases exponentially, and the inference quickly becomes intractable. To overcome the difficulty, the Gaussian mixture reduction, which approximates a high order Gaussian mixture by one with a lower order, can be used. Existing methods such as the clustering-based approaches are renowned for their satisfactory performance and computationally efficiency. However, they have unknown convergence and optimal targets. We propose a novel optimization-based Gaussian mixture reduction method. We develop a majorization-minimization algorithm for its numerical computation and establish its theoretical convergence under general conditions. We show many existing clustering-based methods are special cases of ours, thus bridging the gap between optimization-based and clustering-based methods. The unified framework allows users to choose the most suitable cost function to achieve superior performance in their specific application. We demonstrate the efficiency and effectiveness of the proposed method through extensive empirical experiments.

View on arXiv
Comments on this paper