Processing math: 100%

Rylan Schaeffer

Logo
Resume
Research
Learning
Blog
Teaching
Jokes
Kernel Papers


“Through my blue fingers, pink grains are falling, haphazard, random, a disorganized stream of silicone that seems pregnant with the possibility of every conceivable shape... But this is illusion. Things have their shape in time, not space alone.“

Estimators

An estimator is a rule for calculating some estimate of a desired quantity using data. The rule is the estimator, the thing to be estimated is the estimand and the result of applying the rule to data is the estimate.

Properties

Let data xp(x), θ be the estimand and ˆθ(˙) be an estimator based on x. We define the following properties for the estimator:

Bias

B(ˆθ)=Ep(x)[ˆθ(x)]θ

Variance

Vp(x)=Ep(x)[(ˆθ(x)Ep(x)[ˆθ(x)])2]

Mean Squared Error

MSE(ˆθ(x))=Ep(x)[(ˆθ(x)θ)2]

Bias-Variance Tradeoff

One commonly referenced topic in introductory ML courses is a so-called “bias-variance” tradeoff, which is the fact that the MSE is exactly the sum of the variance plus the bias squared; consequently, for a given MSE, attempting to minimize the variance of an estimator necessary introduces bias and vice versa. To show why this is, we drop x and p(x) for brevity:

MSE(ˆθ)=Ep(x)[(ˆθθ)2]=E[(ˆθ+E[ˆθ]E[ˆθ]θ)2]=E[(ˆθ+E[ˆθ])2]+2E[(ˆθ+E[ˆθ])(E[ˆθ]θ)]E[ˆθ]θ)2]=Vp(x)+B(ˆθ)2

Estimator Desiderata

Consistent

An estimator is consistent if

Unbiased

An estimator is said to be unbiased if the bias B(ˆθ)=0. An unbiased estimator means that on average, the estimate ˆθ(x) equals the estimand θ.

Efficient

A sequence of

Minimal Variance