def. Statistic. let observable random variables [= data of an experiment]. then statistic is:
- i.e. is a real-valued function
- cannot contain unknown variables
def. Estimator [= point estimate] is a statistic used to estimate the parameter of the model we think the data is showing. Note the following notation convention:
- Assume as an r.v. of an experiment, whose model includes parameter .
- To estimate ground truth parameter , we can use an estimator r.v.
- A specific estimate for a particular observed value is denoted
- An estimator has to be a function of known variables & data only.
- , NOT ← This is MSE
How Good is Your Estimator?
- Accuracy is higher. Increased as Bias (Statistics) is decreased
- Precision is higher. Increased as Variance is decreased
- Efficiency (Statistics) is higher. If estimators have the same accuracy, but then the former is more efficient than the latter.
- Consistency.
- Mean Squared Error is lower.
- Likelihood (Statistics) is higher.
→ In general, making sure to reduce bias of estimators is important. Note that:
- If you can write down what the bias is mathematically [= characterize the bias], then you can make a new estimator that doesn’t have the bias.
- Bias usually decreases as the data points increase
Example
let and let estimator where
- are weights that sum to 1. [= weighted average]
- is estimating . is known.
How accurate is ? [=what is the bias?]
How precise is ? What are the best ?
→ Thus is minimized when .