ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1504.01227
22
111

Chebyshev polynomials, moment matching, and optimal estimation of the unseen

6 April 2015
Yihong Wu
Pengkun Yang
ArXivPDFHTML
Abstract

We consider the problem of estimating the support size of a discrete distribution whose minimum non-zero mass is at least 1k \frac{1}{k}k1​. Under the independent sampling model, we show that the sample complexity, i.e., the minimal sample size to achieve an additive error of ϵk\epsilon kϵk with probability at least 0.1 is within universal constant factors of klog⁡klog⁡21ϵ \frac{k}{\log k}\log^2\frac{1}{\epsilon} logkk​log2ϵ1​, which improves the state-of-the-art result of kϵ2log⁡k \frac{k}{\epsilon^2 \log k} ϵ2logkk​ in \cite{VV13}. Similar characterization of the minimax risk is also obtained. Our procedure is a linear estimator based on the Chebyshev polynomial and its approximation-theoretic properties, which can be evaluated in O(n+log⁡2k)O(n+\log^2 k)O(n+log2k) time and attains the sample complexity within a factor of six asymptotically. The superiority of the proposed estimator in terms of accuracy, computational efficiency and scalability is demonstrated in a variety of synthetic and real datasets.

View on arXiv
Comments on this paper