Greedy Matching and Propensity Score Matching
2022-07-09

Greedy (Nearest-neighbor) Matching

Pair Matching

Steps:

  1. Randomly order list of treated subjects and control subjects

  2. Start with the first treated subject. Match to the control with the smallest distance (this is greedy).

  3. Remove the matched control from the list of available matches.

  4. Move on to the next treated subject. Match to the control with the smallest distance.

  5. Repeat steps 3 and 4 until you have matched all treated subjects.

Advantage Disadvantage
1. Intuitive
2. Computationally fast
- Involves a series of simple algorithms (identifying minimum distance)
- Fast even for large data sets
- R package: MatchIt
1. Not invariant to initial order of list
2. Not optimal
- Always taking the smallest distance match does not minimize total distance
-Can lead to some bad matches (think of step3 above)

Many-to-one Matching

For k:1 matching: After everyone has 1 match, go through the list again and find 2nd matches. Repeat until k matches.

Advantage Disadvantage
End up using more of the controls and a large sample size. It suggests a more efficient estimate of the causal effect. Adding a new treated subject to your data set would be more of an efficiency gain than adding another control to a matched set. So there is some power gain, some efficiency gain, but it's not very large. The many-to-one matching should lead to a little more bias, but smaller variance.

Caliper

A bad match can be defined using a caliper - maximum acceptable distance.

  • Only match a treated subject if the best match has a distance less than the caliper

  • Recall positivity assumption: Probability of each treatment given X should be non-zero, i.e., .

    • If no matches within the caliper, it is a sign that the positivity assumption would be violated

    • Excluding these subjects makes the assumption more realistic

    • Drawback: Hard to define

Optimal Matching

  • Minimize global distance (i.e., total distance)

    • Greedy matching is not necessarily optimal and usually is not in terms of minimizing the total distance. Because there might be times when you want to save a match for a later subject and accept a slightly less good match now.
  • Computationally demanding

    • R packages: optmatch, rcbalance

Balance Assessment

  • Covariate Balance

    • Two sample t-tests for continuous variables or Chi-sqaure tests for discrete variables

    • Drawback: p-values are dependent on the sample size. Particularly, small differences in means can be significant with a small p-value if the sample size is large. This is what we don't want.

    • Standardized (mean) differences

      Values Balance
      <0.1 Adequate balance
      0.1-0.2 Not too alarming
      >0.2 Serious imbalance

Randomization Tests

Also known as Permutation Tests and Exact Tests

Test Steps:

  1. Compute the test statistic from observed data

  2. Assume the null hypothesis of no treatment effect is true

  3. Randomly permute treatment assignment within pairs and re-compute test statistic

  4. Repeat many times and see how unusual the observed statistic is

Notes:

  • When calculating a p-value, the intuition is to calculate the probability of something as extreme or more extreme than our particular value

  • This test is equivalent to the McNemar test for paired binary/discrete data and equivalent to paired t-test for paired continuous data

Sensitivity Analysis

  • Overt bias could occur if there was imbalance on observed covariates

  • There is no guarantee that matching will result in balance on variables that we did not match on (including unobserved variabels)

Motivation

If there is hidden bias, how severe it would have to be to change conclusions.

  • Let and be the probability that person and receive treatment

  • Suppose person and are perfectly matched so that their observed covariates and are the same.

  • Inequality

    where is odds ratio.

    • If , then , suggesting no hidden bias

    • implies hidden bias

    • is the odds of treatment for person

Propensity Score

The propensity score is the probability of receiving treatment rather than control, given covariates X:

Balancing Score

Formally,

where represents the distributions of covariates at every value and the propensity score function generates the value of .

Implication: If we match according to the propensity score, we should achieve balance.

Estimated Propensity Score

Estimate to obtain

  • The outcome is A, which is binary

  • Estimate the propensity score using logistic regression

    1. Fit a logistic regression model: outcome A and covariates X

    2. From that model, get the predicted probability (fitted value) for each subject, which is the estimated propensity score,

In the randomized control trials, the propensity score is 0.5 by design, since given X subjects are randomly assigned to the treatment or the control group.

Propensity Score Matching

The propensity score is a scalar. The matching problem is simplified, in that we are only matching on one variable.

Trimming Tails

Remove subjects who have extreme values of the propensity scores

  • Control subjects whose propensity score is less than the minimum in the treatment group

  • Treated subjects whose propensity score is greater than the maximum in the control group

Trimming the tails makes the positivity assumption more reasonable and prevents extrapolation.

Matching

Compute a distance between the propensity score for each treated subject with every control. Then use the nearest neighbor (i.e., greedy mathcing) or optimal matching as before.

In practice, logit (log-odds) of the propensity score is often used, rather than the propensity score itself.

  • The propensity score is bounded between 0 and 1, making many values similar

  • Logit of the propensity score is unbounded -- this transformation essentially stretches the distribution, while preserving ranks.

  • Match on rather than

Caliper

In practice, a common choice for a caliper (i.e., the maximum distance we are willing to tolerate in matching) is 0.2 times the standard deviation of the logit of the propensity score.

  1. Estimate the propensity score using logistic regression

  2. logit-transform the propensity score

  3. Take the standard deviation of this transformed variable

  4. Set the caliper to 0.2 times the value from step 3

A smaller caliper suggests less bias but more variance.

Post-matching

The outcome analysis methods can be the same as those that would be used if matching directly on covariates. For example,

  • Randomization tests

  • Conditional logistic regression, GEE, Stratified Cox Model