ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.16008
29
0

Exact Recovery of Sparse Binary Vectors from Generalized Linear Measurements

21 February 2025
Arya Mazumdar
Neha Sangwan
    MQ
ArXivPDFHTML
Abstract

We consider the problem of exact recovery of a kkk-sparse binary vector from generalized linear measurements (such as logistic regression). We analyze the linear estimation algorithm (Plan, Vershynin, Yudovina, 2017), and also show information theoretic lower bounds on the number of required measurements. As a consequence of our results, for noisy one bit quantized linear measurements (1bCSbinary\mathsf{1bCSbinary}1bCSbinary), we obtain a sample complexity of O((k+σ2)log⁡n)O((k+\sigma^2)\log{n})O((k+σ2)logn), where σ2\sigma^2σ2 is the noise variance. This is shown to be optimal due to the information theoretic lower bound. We also obtain tight sample complexity characterization for logistic regression.Since 1bCSbinary\mathsf{1bCSbinary}1bCSbinary is a strictly harder problem than noisy linear measurements (SparseLinearReg\mathsf{SparseLinearReg}SparseLinearReg) because of added quantization, the same sample complexity is achievable for SparseLinearReg\mathsf{SparseLinearReg}SparseLinearReg. While this sample complexity can be obtained via the popular lasso algorithm, linear estimation is computationally more efficient. Our lower bound holds for any set of measurements for SparseLinearReg\mathsf{SparseLinearReg}SparseLinearReg, (similar bound was known for Gaussian measurement matrices) and is closely matched by the maximum-likelihood upper bound. For SparseLinearReg\mathsf{SparseLinearReg}SparseLinearReg, it was conjectured in Gamarnik and Zadik, 2017 that there is a statistical-computational gap and the number of measurements should be at least (2k+σ2)log⁡n(2k+\sigma^2)\log{n}(2k+σ2)logn for efficient algorithms to exist. It is worth noting that our results imply that there is no such statistical-computational gap for 1bCSbinary\mathsf{1bCSbinary}1bCSbinary and logistic regression.

View on arXiv
@article{mazumdar2025_2502.16008,
  title={ Exact Recovery of Sparse Binary Vectors from Generalized Linear Measurements },
  author={ Arya Mazumdar and Neha Sangwan },
  journal={arXiv preprint arXiv:2502.16008},
  year={ 2025 }
}
Comments on this paper