ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2002.04232
216
4
v1v2 (latest)

Efficiently Learning and Sampling Interventional Distributions from Observations

11 February 2020
Arnab Bhattacharyya
Sutanu Gayen
S. Kandasamy
Ashwin Maran
N. V. Vinodchandran
    CML
ArXiv (abs)PDFHTML
Abstract

We study the problem of efficiently estimating the effect of an intervention on a single variable using observational samples in a causal Bayesian network. Our goal is to give algorithms that are efficient in both time and sample complexity in a non-parametric setting. Tian and Pearl (AAAI `02) have exactly characterized the class of causal graphs for which causal effects of atomic interventions can be identified from observational data. We make their result quantitative. Suppose P is a causal model on a set V of n observable variables with respect to a given causal graph G with observable distribution PPP. Let PxP_xPx​ denote the interventional distribution over the observables with respect to an intervention of a designated variable X with x. We show that assuming that G has bounded in-degree, bounded c-components, and that the observational distribution is identifiable and satisfies certain strong positivity condition: 1. [Evaluation] There is an algorithm that outputs with probability 2/32/32/3 an evaluator for a distribution P′P'P′ that satisfies dtv(Px,P′)≤ϵd_{tv}(P_x, P') \leq \epsilondtv​(Px​,P′)≤ϵ using m=O~(nϵ−2)m=\tilde{O}(n\epsilon^{-2})m=O~(nϵ−2) samples from PPP and O(mn)O(mn)O(mn) time. The evaluator can return in O(n)O(n)O(n) time the probability P′(v)P'(v)P′(v) for any assignment vvv to VVV. 2. [Generation] There is an algorithm that outputs with probability 2/32/32/3 a sampler for a distribution P^\hat{P}P^ that satisfies dtv(Px,P^)≤ϵd_{tv}(P_x, \hat{P}) \leq \epsilondtv​(Px​,P^)≤ϵ using m=O~(nϵ−2)m=\tilde{O}(n\epsilon^{-2})m=O~(nϵ−2) samples from PPP and O(mn)O(mn)O(mn) time. The sampler returns an iid sample from P^\hat{P}P^ with probability 1−δ1-\delta1−δ in O(nϵ−1log⁡δ−1)O(n\epsilon^{-1} \log\delta^{-1})O(nϵ−1logδ−1) time. We extend our techniques to estimate marginals Px∣YP_x|_YPx​∣Y​ over a given Y⊂VY \subset VY⊂V of interest. We also show lower bounds for the sample complexity showing that our sample complexity has optimal dependence on the parameters n and ϵ\epsilonϵ as well as the strong positivity parameter.

View on arXiv
Comments on this paper