Propensity Score Matching And Doubly Robust Estimation In Causal Inference

Propensity score matching and doubly robust estimation are two powerful techniques used in causal inference to address selection bias. Propensity score matching aims to create a synthetic control group for a treatment group by matching individuals based on their propensity to receive the treatment. Doubly robust estimation, on the other hand, combines a propensity score-based model with a non-parametric model to estimate the causal effect of a treatment. By leveraging both propensity score matching and doubly robust estimation, researchers can obtain more robust and accurate estimates in observational studies.

Propensity Score Matching with Doubly Robust Estimation

Propensity score matching (PSM) is a popular method used to estimate the causal effect of a treatment when randomized experiments are not feasible. Doubly robust (DR) estimators are estimators that combine model predictions and propensity scores to improve the accuracy of the estimated treatment effect.

Doubly Robust Estimator Structure

The doubly robust propensity score matching estimator is given by:

E[Y(1) - Y(0) | X] = E[Y(1) - Y(0) | X, R] - E[Y(1) - Y(0) | X] * E[R | X]

Where:

  • Y(1) and Y(0) are the potential outcomes for the treatment and control groups, respectively
  • X are the observed covariates
  • R is the propensity score

Implementation

  1. Estimate the propensity score: Use a logistic regression model to predict the probability of treatment assignment given the observed covariates.
  2. Match treated and control units: Pair each treated unit with a control unit with a similar propensity score. Matching can be done using methods such as nearest neighbor matching or kernel matching.
  3. Estimate the treatment effect using a regression model: Regress the outcome variable on the treatment status and the propensity score for the matched pairs.

Advantages of Doubly Robust Estimation

  • Reduces bias due to incorrect model specification
  • More efficient than inverse probability weighting (IPW)
  • Can handle non-ignorable missing data

Table: Matching Methods with Doubly Robust Estimation

Matching Method Model
Nearest neighbor matching Regression
Kernel matching Non-parametric
Stratification None

Limitations

  • Requires strong overlap in propensity scores between the treatment and control groups
  • Can be sensitive to model misspecification

Question 1:

How does propensity score matching become doubly robust when combined with an outcome regression model?

Answer:

Propensity score matching (PSM) becomes doubly robust when an outcome regression model (ORM) is introduced because it creates two sets of estimates: one based on PSM alone and one based on PSM combined with the ORM. The doubly robust property ensures that either of these estimates will be consistent (i.e., unbiased) if either the PSM or the ORM is correctly specified.

Question 2:

What is the key difference between inverse probability of treatment weighting (IPTW) and propensity score matching (PSM) in causal inference?

Answer:

IPTW and PSM are both methods for estimating causal effects by creating a treatment and control group that are similar in observed characteristics. The key difference is that IPTW uses weights to adjust for differences between the groups, while PSM creates a matched sample by selecting controls who are similar to the treated individuals on a single balancing score called the propensity score.

Question 3:

How can propensity score stratification (PSS) be used to improve the precision of propensity score matching (PSM)?

Answer:

PSS divides the sample into strata based on their propensity scores and performs separate PSM within each stratum. This reduces the heterogeneity within the strata, which leads to more precise matching and, consequently, more efficient estimation of causal effects.

Well, there you have it, folks! We’ve covered the basics of propensity score matching doubly robust. It’s a powerful technique that can help you make more accurate causal inferences from observational data. Of course, there’s a lot more to learn, but this should give you a good foundation. Thanks for reading! If you have any questions, feel free to leave a comment below. And be sure to check back soon for more data science goodness.

Leave a Comment