Bias correction, estimating equations, and TMLE all give efficient estimators that are asymptotically equivalent. Precisely, where is the efficient influence function.
In large samples, remember that we have by the CLT applied to the above that where . We can therefore do inference by computing , which justifies building 95% confidence intervals of the form . Since is approximately , a Z-test on this statistic is appropriate to obtain a p-value.
A brief proof that (i.e. that this estimate is consistent) under the same assumptions we've been operating under is given below and relies on techniques similar to those discussed in the section on naive plug-in estimation.
Proof:
We are trying to show that . Decompose the difference as
The first term is an empirical process term that we know is and therefore as long as is -consistent and we have either used sample splitting or satisfied some Donsker conditions. The -consistency of is indeed guaranteed by the presumed -consistency of as long as is not permitted to blow up (e.g. we bound , as previously required for efficiency of ). The proof of this is an application of the bounded convergence theorem.
The second term times converges in distribution to by the central limit theorem and is thus and thus (think: if this term "blown up" by stabilizes to a normal with finite variance, then if we take away the it must collapse down to its mean, which is 0).
The last term is bounded by according to Cauchy-Schwarz (again assuming doesn't blow up) and therefore by the -consistency of is also .
Since all three terms are , their sum is as well, which is what we wanted.
It's also possible to construct alternative estimators for the variance (e.g. bootstrap) or even targeted, efficient estimators using the same methodologies used to estimate in the first place! However, for the most part, the sample variance of the estimated influence function () suffices.