ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2207.00956
23
8

Tricking the Hashing Trick: A Tight Lower Bound on the Robustness of CountSketch to Adaptive Inputs

3 July 2022
Edith Cohen
Jelani Nelson
Tamas Sarlos
Uri Stemmer
    AAML
ArXivPDFHTML
Abstract

CountSketch and Feature Hashing (the "hashing trick") are popular randomized dimensionality reduction methods that support recovery of ℓ2\ell_2ℓ2​-heavy hitters (keys iii where vi2>ϵ∥v∥22v_i^2 > \epsilon \|\boldsymbol{v}\|_2^2vi2​>ϵ∥v∥22​) and approximate inner products. When the inputs are {\em not adaptive} (do not depend on prior outputs), classic estimators applied to a sketch of size O(ℓ/ϵ)O(\ell/\epsilon)O(ℓ/ϵ) are accurate for a number of queries that is exponential in ℓ\ellℓ. When inputs are adaptive, however, an adversarial input can be constructed after O(ℓ)O(\ell)O(ℓ) queries with the classic estimator and the best known robust estimator only supports O~(ℓ2)\tilde{O}(\ell^2)O~(ℓ2) queries. In this work we show that this quadratic dependence is in a sense inherent: We design an attack that after O(ℓ2)O(\ell^2)O(ℓ2) queries produces an adversarial input vector whose sketch is highly biased. Our attack uses "natural" non-adaptive inputs (only the final adversarial input is chosen adaptively) and universally applies with any correct estimator, including one that is unknown to the attacker. In that, we expose inherent vulnerability of this fundamental method.

View on arXiv
Comments on this paper