Greedy (Nearest-neighbor) Matching
Pair Matching
Steps:
Randomly order list of treated subjects and control subjects
Start with the first treated subject. Match to the control with the smallest distance (this is greedy).
Remove the matched control from the list of available matches.
Move on to the next treated subject. Match to the control with the smallest distance.
Repeat steps 3 and 4 until you have matched all treated subjects.
Advantage | Disadvantage |
---|---|
1. Intuitive 2. Computationally fast - Involves a series of simple algorithms (identifying minimum distance) - Fast even for large data sets - R package: MatchIt |
1. Not invariant to initial order of list 2. Not optimal - Always taking the smallest distance match does not minimize total distance -Can lead to some bad matches (think of step3 above) |
Many-to-one Matching
For k:1 matching: After everyone has 1 match, go through the list again and find 2nd matches. Repeat until k matches.
Advantage | Disadvantage |
---|---|
End up using more of the controls and a large sample size. It suggests a more efficient estimate of the causal effect. | Adding a new treated subject to your data set would be more of an efficiency gain than adding another control to a matched set. So there is some power gain, some efficiency gain, but it's not very large. The many-to-one matching should lead to a little more bias, but smaller variance. |
Caliper
A bad match can be defined using a caliper - maximum acceptable distance.
Only match a treated subject if the best match has a distance less than the caliper
Recall positivity assumption: Probability of each treatment given X should be non-zero, i.e.,
. If no matches within the caliper, it is a sign that the positivity assumption would be violated
Excluding these subjects makes the assumption more realistic
Drawback: Hard to define
Optimal Matching
Minimize global distance (i.e., total distance)
- Greedy matching is not necessarily optimal and usually is not in terms of minimizing the total distance. Because there might be times when you want to save a match for a later subject and accept a slightly less good match now.
Computationally demanding
- R packages:
optmatch
,rcbalance
- R packages:
Balance Assessment
Covariate Balance
Two sample t-tests for continuous variables or Chi-sqaure tests for discrete variables
Two sample t-test (equal variance)
where
and are sample sizes, is the population deviation such that When
, reduces to . Two sample t-test (unequal variance)
Chi-Square Test of Independence | Introduction to Statistics | JMP
Chi-Square Goodness of Fit Test | Introduction to Statistics | JMP
Drawback: p-values are dependent on the sample size. Particularly, small differences in means can be significant with a small p-value if the sample size is large. This is what we don't want.
Standardized (mean) differences
Values Balance <0.1 Adequate balance 0.1-0.2 Not too alarming >0.2 Serious imbalance
Randomization Tests
Also known as Permutation Tests and Exact Tests
Test Steps:
Compute the test statistic from observed data
Assume the null hypothesis of no treatment effect is true
Randomly permute treatment assignment within pairs and re-compute test statistic
Repeat many times and see how unusual the observed statistic is
Notes:
When calculating a p-value, the intuition is to calculate the probability of something as extreme or more extreme than our particular value
This test is equivalent to the McNemar test for paired binary/discrete data and equivalent to paired t-test for paired continuous data
Sensitivity Analysis
Overt bias could occur if there was imbalance on observed covariates
There is no guarantee that matching will result in balance on variables that we did not match on (including unobserved variabels)
If these unobserved variables are confounders. then we have hidden bias and ignorability assumption violated (i.e.,
) Understanding of Ignorability:
where if treated, 0 otherwise. Ignorability says certainly affects observed outcome by defination but does not relate to the values of and .
Motivation
If there is hidden bias, how severe it would have to be to change conclusions.
Let
and be the probability that person and receive treatment Suppose person
and are perfectly matched so that their observed covariates and are the same. Inequality
where
is odds ratio. If
, then , suggesting no hidden bias implies hidden bias is the odds of treatment for person
Propensity Score
The propensity score is the probability of receiving treatment rather than control, given covariates X:
Balancing Score
Formally,
where
Implication: If we match according to the propensity score, we should achieve balance.
Estimated Propensity Score
Estimate
The outcome is A, which is binary
Estimate the propensity score using logistic regression
Fit a logistic regression model: outcome A and covariates X
From that model, get the predicted probability (fitted value) for each subject, which is the estimated propensity score,
In the randomized control trials, the propensity score is 0.5 by design, since given X subjects are randomly assigned to the treatment or the control group.
Propensity Score Matching
The propensity score is a scalar. The matching problem is simplified, in that we are only matching on one variable.
Trimming Tails
Remove subjects who have extreme values of the propensity scores
Control subjects whose propensity score is less than the minimum in the treatment group
Treated subjects whose propensity score is greater than the maximum in the control group
Trimming the tails makes the positivity assumption more reasonable and prevents extrapolation.
Matching
Compute a distance between the propensity score for each treated subject with every control. Then use the nearest neighbor (i.e., greedy mathcing) or optimal matching as before.
In practice, logit (log-odds) of the propensity score is often used, rather than the propensity score itself.
The propensity score is bounded between 0 and 1, making many values similar
Logit of the propensity score is unbounded -- this transformation essentially stretches the distribution, while preserving ranks.
Match on
rather than
Caliper
In practice, a common choice for a caliper (i.e., the maximum distance we are willing to tolerate in matching) is 0.2 times the standard deviation of the logit of the propensity score.
Estimate the propensity score using logistic regression
logit-transform the propensity score
Take the standard deviation of this transformed variable
Set the caliper to 0.2 times the value from step 3
A smaller caliper suggests less bias but more variance.
Post-matching
The outcome analysis methods can be the same as those that would be used if matching directly on covariates. For example,
Randomization tests
Conditional logistic regression, GEE, Stratified Cox Model
View / Make Comments