ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.00413
13
4

Fast Generating A Large Number of Gumbel-Max Variables

2 February 2020
Yiyan Qi
P. Wang
Yuanming Zhang
Junzhou Zhao
Guangjian Tian
X. Guan
ArXivPDFHTML
Abstract

The well-known Gumbel-Max Trick for sampling elements from a categorical distribution (or more generally a nonnegative vector) and its variants have been widely used in areas such as machine learning and information retrieval. To sample a random element iii (or a Gumbel-Max variable iii) in proportion to its positive weight viv_ivi​, the Gumbel-Max Trick first computes a Gumbel random variable gig_igi​ for each positive weight element iii, and then samples the element iii with the largest value of gi+ln⁡vig_i+\ln v_igi​+lnvi​. Recently, applications including similarity estimation and graph embedding require to generate kkk independent Gumbel-Max variables from high dimensional vectors. However, it is computationally expensive for a large kkk (e.g., hundreds or even thousands) when using the traditional Gumbel-Max Trick. To solve this problem, we propose a novel algorithm, \emph{FastGM}, that reduces the time complexity from O(kn+)O(kn^+)O(kn+) to O(kln⁡k+n+)O(k \ln k + n^+)O(klnk+n+), where n+n^+n+ is the number of positive elements in the vector of interest. Instead of computing kkk independent Gumbel random variables directly, we find that there exists a technique to generate these variables in descending order. Using this technique, our method FastGM computes variables gi+ln⁡vig_i+\ln v_igi​+lnvi​ for all positive elements iii in descending order. As a result, FastGM significantly reduces the computation time because we can stop the procedure of Gumbel random variables computing for many elements especially for those with small weights. Experiments on a variety of real-world datasets show that FastGM is orders of magnitude faster than state-of-the-art methods without sacrificing accuracy and incurring additional expenses.

View on arXiv
Comments on this paper