ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2011.09986
12
0

Estimation of Shortest Path Covariance Matrices

19 November 2020
R. Maity
Cameron Musco
ArXivPDFHTML
Abstract

We study the sample complexity of estimating the covariance matrix Σ∈Rd×d\mathbf{\Sigma} \in \mathbb{R}^{d\times d}Σ∈Rd×d of a distribution D\mathcal DD over Rd\mathbb{R}^dRd given independent samples, under the assumption that Σ\mathbf{\Sigma}Σ is graph-structured. In particular, we focus on shortest path covariance matrices, where the covariance between any two measurements is determined by the shortest path distance in an underlying graph with ddd nodes. Such matrices generalize Toeplitz and circulant covariance matrices and are widely applied in signal processing applications, where the covariance between two measurements depends on the (shortest path) distance between them in time or space. We focus on minimizing both the vector sample complexity: the number of samples drawn from D\mathcal{D}D and the entry sample complexity: the number of entries read in each sample. The entry sample complexity corresponds to measurement equipment costs in signal processing applications. We give a very simple algorithm for estimating Σ\mathbf{\Sigma}Σ up to spectral norm error ϵ∥Σ∥2\epsilon \left\|\mathbf{\Sigma}\right\|_2ϵ∥Σ∥2​ using just O(D)O(\sqrt{D})O(D​) entry sample complexity and O~(r2/ϵ2)\tilde O(r^2/\epsilon^2)O~(r2/ϵ2) vector sample complexity, where DDD is the diameter of the underlying graph and r≤dr \le dr≤d is the rank of Σ\mathbf{\Sigma}Σ. Our method is based on extending the widely applied idea of sparse rulers for Toeplitz covariance estimation to the graph setting. In the special case when Σ\mathbf{\Sigma}Σ is a low-rank Toeplitz matrix, our result matches the state-of-the-art, with a far simpler proof. We also give an information theoretic lower bound matching our upper bound up to a factor DDD and discuss some directions towards closing this gap.

View on arXiv
Comments on this paper