Inverse probability weighting is a statistical technique for estimating quantities related to a population other than the one from which the data was collected.
[1] There may be prohibitive factors barring researchers from directly sampling from the target population such as cost, time, or ethical concerns.
[2] A solution to this problem is to use an alternate design strategy, e.g. stratified sampling.
Weighting, when correctly applied, can potentially improve the efficiency and reduce the bias of unweighted estimators.
[3] When the sampling probability is known, from which the sampling population is drawn from the target population, then the inverse of this probability is used to weight the observations.
This approach has been generalized to many aspects of statistics under various frameworks.
These applications codified the theory of other statistics and estimators such as marginal structural models, the standardized mortality ratio, and the EM algorithm for coarsened or aggregate data.
Inverse probability weighting is also used to account for missing data when subjects with missing data cannot be included in the primary analysis.
[4] With an estimate of the sampling probability, or the probability that the factor would be measured in another measurement, inverse probability weighting can be used to inflate the weight for subjects who are under-represented due to a large degree of missing data.
The inverse probability weighting estimator can be used to demonstrate causality when the researcher cannot conduct a controlled experiment but has observed data to model.
Because it is assumed that the treatment is not randomly assigned, the goal is to estimate the counterfactual or potential outcome if all subjects in population were assigned either treatment.
drawn i.i.d (independent and identically distributed) from unknown distribution P, where The goal is to estimate the potential outcome,
, that would be observed if the subject were assigned treatment
Then compare the mean outcome if all patients in the population were assigned either treatment:
With the mean of each treatment group computed, a statistical t-test or ANOVA test can be used to judge difference between group means and determine statistical significance of treatment effect.
Under the assumptions (A1)-(A3), we will derive the following identities
For the second equality, first use the iterated expectation to write
In such instances, the IPWE can be dominated by a small number of subjects with large weights.
To address this issue, a smoothed IPW estimator using Rao-Blackwellization has been proposed, which reduces the variance of IPWE by up to 7-fold and helps protect the estimator from model misspecification.
It is therefore a 'doubly robust' method in that it only requires either the propensity or outcome model to be correctly specified but not both.
This method augments the IPWE to reduce variability and improve estimate efficiency.
This model holds the same assumptions as the Inverse Probability Weighted Estimator (IPWE).
With the following notations: The later rearrangement of the formula helps reveal the underlying idea: our estimator is based on the average predicted outcome using the model (i.e.:
We can correct this potential bias by adding the extra term of the average residuals of the model (Q) from the true value of the outcome (Y) (i.e.:
Because we have missing values of Y, we give weights to inflate the relative importance of each residual (these weights are based on the inverse propensity, a.k.a.
probability, of seeing each subject observations) (see page 10 in [7]).
The "doubly robust" benefit of such an estimator comes from the fact that it's sufficient for one of the two models to be correctly specified, for the estimator to be unbiased (either
While if the model is biased, but the weighting model is well specified, then the bias will be well estimated (And corrected for) by the weighted average residuals.
[7][8][9] The bias of the doubly robust estimators is called a second-order bias, and it depends on the product of the difference
This property allows us, when having a "large enough" sample size, to lower the overall bias of doubly robust estimators by using machine learning estimators (instead of parametric models).