“Through my blue fingers, pink grains are falling, haphazard, random, a disorganized stream of silicone that seems pregnant with the possibility of every conceivable shape... But this is illusion. Things have their shape in time, not space alone.“
An estimator is a rule for calculating some estimate of a desired quantity using data. The rule is the estimator, the thing to be estimated is the estimand and the result of applying the rule to data is the estimate.
Let data x∼p(x), θ be the estimand and ˆθ(˙) be an estimator based on x. We define the following properties for the estimator:
One commonly referenced topic in introductory ML courses is a so-called “bias-variance” tradeoff, which is the fact that the MSE is exactly the sum of the variance plus the bias squared; consequently, for a given MSE, attempting to minimize the variance of an estimator necessary introduces bias and vice versa. To show why this is, we drop x and p(x) for brevity:
MSE(ˆθ)=Ep(x)[(ˆθ−θ)2]=E[(ˆθ+E[ˆθ]−E[ˆθ]−θ)2]=E[(ˆθ+E[ˆθ])2]+2E[(ˆθ+E[ˆθ])(E[ˆθ]−θ)]−E[ˆθ]−θ)2]=Vp(x)+B(ˆθ)2An estimator is consistent if
An estimator is said to be unbiased if the bias B(ˆθ)=0. An unbiased estimator means that on average, the estimate ˆθ(x) equals the estimand θ.
A sequence of