ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2305.04712
50
3

High-Dimensional Smoothed Entropy Estimation via Dimensionality Reduction

8 May 2023
Kristjan Greenewald
Brian Kingsbury
Yuancheng Yu
ArXivPDFHTML
Abstract

We study the problem of overcoming exponential sample complexity in differential entropy estimation under Gaussian convolutions. Specifically, we consider the estimation of the differential entropy h(X+Z)h(X+Z)h(X+Z) via nnn independently and identically distributed samples of XXX, where XXX and ZZZ are independent DDD-dimensional random variables with XXX sub-Gaussian with bounded second moment and Z∼N(0,σ2ID)Z\sim\mathcal{N}(0,\sigma^2I_D)Z∼N(0,σ2ID​). Under the absolute-error loss, the above problem has a parametric estimation rate of cDn\frac{c^D}{\sqrt{n}}n​cD​, which is exponential in data dimension DDD and often problematic for applications. We overcome this exponential sample complexity by projecting XXX to a low-dimensional space via principal component analysis (PCA) before the entropy estimation, and show that the asymptotic error overhead vanishes as the unexplained variance of the PCA vanishes. This implies near-optimal performance for inherently low-dimensional structures embedded in high-dimensional spaces, including hidden-layer outputs of deep neural networks (DNN), which can be used to estimate mutual information (MI) in DNNs. We provide numerical results verifying the performance of our PCA approach on Gaussian and spiral data. We also apply our method to analysis of information flow through neural network layers (c.f. information bottleneck), with results measuring mutual information in a noisy fully connected network and a noisy convolutional neural network (CNN) for MNIST classification.

View on arXiv
Comments on this paper