Formally speaking, an estimator Tn of parameter θ is said to be weakly consistent, if it converges in probability to the true value of the parameter:1
i.e. if, for all ε > 0
An estimator Tn of parameter θ is said to be strongly consistent, if it converges almost surely to the true value of the parameter:
A more rigorous definition takes into account the fact that θ is actually unknown, and thus, the convergence in probability must take place for every possible value of this parameter. Suppose {pθ: θ ∈ Θ} is a family of distributions (the parametric model), and Xθ = {X1, X2, … : Xi ~ pθ} is an infinite sample from the distribution pθ. Let { Tn(Xθ) } be a sequence of estimators for some parameter g(θ). Usually, Tn will be based on the first n observations of a sample. Then this sequence {Tn} is said to be (weakly) consistent if 2
This definition uses g(θ) instead of simply θ, because often one is interested in estimating a certain function or a sub-vector of the underlying parameter. In the next example, we estimate the location parameter of the model, but not the scale:
Suppose one has a sequence of statistically independent observations {X1, X2, ...} from a normal N(μ, σ2) distribution. To estimate μ based on the first n observations, one can use the sample mean: Tn = (X1 + ... + Xn)/n. This defines a sequence of estimators, indexed by the sample size n.
From the properties of the normal distribution, we know the sampling distribution of this statistic: Tn is itself normally distributed, with mean μ and variance σ2/n. Equivalently, ( T n − μ ) / ( σ / n ) {\displaystyle \scriptstyle (T_{n}-\mu )/(\sigma /{\sqrt {n}})} has a standard normal distribution:
as n tends to infinity, for any fixed ε > 0. Therefore, the sequence Tn of sample means is consistent for the population mean μ (recalling that Φ {\displaystyle \Phi } is the cumulative distribution of the standard normal distribution).
The notion of asymptotic consistency is very close, almost synonymous to the notion of convergence in probability. As such, any theorem, lemma, or property which establishes convergence in probability may be used to prove the consistency. Many such tools exist:
the most common choice for function h being either the absolute value (in which case it is known as Markov inequality), or the quadratic function (respectively Chebyshev's inequality).
An estimator can be unbiased but not consistent. For example, for an iid sample {x1,..., xn} one can use Tn(X) = xn as the estimator of the mean E[X]. Note that here the sampling distribution of Tn is the same as the underlying distribution (for any n, as it ignores all points but the last). So E[Tn(X)] = E[X] for any n, hence it is unbiased, but it does not converge to any value.
However, if a sequence of estimators is unbiased and converges to a value, then it is consistent, as it must converge to the correct value.
Alternatively, an estimator can be biased but consistent. For example, if the mean is estimated by 1 n ∑ x i + 1 n {\displaystyle {1 \over n}\sum x_{i}+{1 \over n}} it is biased, but as n → ∞ {\displaystyle n\rightarrow \infty } , it approaches the correct value, and so it is consistent.
Important examples include the sample variance and sample standard deviation. Without Bessel's correction (that is, when using the sample size n {\displaystyle n} instead of the degrees of freedom n − 1 {\displaystyle n-1} ), these are both negatively biased but consistent estimators. With the correction, the corrected sample variance is unbiased, while the corrected sample standard deviation is still biased, but less so, and both are still consistent: the correction factor converges to 1 as sample size grows.
Here is another example. Let T n {\displaystyle T_{n}} be a sequence of estimators for θ {\displaystyle \theta } .
We can see that T n → p θ {\displaystyle T_{n}{\xrightarrow {p}}\theta } , E [ T n ] = θ + δ {\displaystyle \operatorname {E} [T_{n}]=\theta +\delta } , and the bias does not converge to zero.
Amemiya 1985, Definition 3.4.2. - Amemiya, Takeshi (1985). Advanced Econometrics. Harvard University Press. ISBN 0-674-00560-0. https://archive.org/details/advancedeconomet00amem ↩
Lehman & Casella 1998, p. 332. - Lehmann, E. L.; Casella, G. (1998). Theory of Point Estimation (2nd ed.). Springer. ISBN 0-387-98502-6. ↩
Amemiya 1985, equation (3.2.5). - Amemiya, Takeshi (1985). Advanced Econometrics. Harvard University Press. ISBN 0-674-00560-0. https://archive.org/details/advancedeconomet00amem ↩
Amemiya 1985, Theorem 3.2.6. - Amemiya, Takeshi (1985). Advanced Econometrics. Harvard University Press. ISBN 0-674-00560-0. https://archive.org/details/advancedeconomet00amem ↩
Amemiya 1985, Theorem 3.2.7. - Amemiya, Takeshi (1985). Advanced Econometrics. Harvard University Press. ISBN 0-674-00560-0. https://archive.org/details/advancedeconomet00amem ↩
Newey & McFadden 1994, Chapter 2. - Newey, W. K.; McFadden, D. (1994). "Chapter 36: Large sample estimation and hypothesis testing". In Robert F. Engle; Daniel L. McFadden (eds.). Handbook of Econometrics. Vol. 4. Elsevier Science. ISBN 0-444-88766-0. S2CID 29436457. https://api.semanticscholar.org/CorpusID:29436457 ↩