Validation Slides
Validation Slides
◼ Training
◼ Model selection
◼ Resampling Techniques
◼ Cross Validation
◼ Bootstrap
◼ Jackknife
The holdout method
Split dataset into two groups
◼ Training set: used to train the classifier
◼ Test set: used to estimate the error rate of the trained classifier
The holdout method has two basic drawbacks
◼ In sparse dataset , setting aside a portion of the dataset for testing not feasible
◼ A single train-and-test experiment, the holdout estimate of error rate dependant
on split of test and train
These limitations of the holdout can be overcome with a family of resampling
methods at the expense of higher computational cost
◼ Cross validation
◼ Random subsampling
◼ K-fold cross-validation
◼ Leave-one-out cross-validation
◼ Bootstrap
◼ Jackknife
Random subsampling
Small k
◼ The number of experiments and, therefore, computation time are reduced
◼ The variance of the estimator will be small
◼ The bias of the estimator will be large (conservative or larger than the true error rate)
◼ Variance Estimate
Linear approximation of bootstrap, only works well for linear statistics (e.g.,
mean). It fails to give accurate estimation for non-smooth (e.g., median) and
nonlinear (e.g., correlation coefficient) cases
Bias and Variance of a Statistical Estimate
Problem: estimate parameter 𝛼 of unknown distribution 𝐺
◼ To emphasize the fact that 𝛼 concerns 𝐺, we write 𝛼(𝐺)
Solution
◼ We collect 𝑁 examples 𝑋={𝑥1 ,𝑥2 …𝑥𝑁} from 𝐺
◼ 𝑋 defines a discrete distribution 𝐺’ with mass 1/𝑁 at each example
◼ We compute the statistic 𝛼’=𝛼(𝐺’) as an estimator of 𝛼(𝐺)
◼ e.g., 𝛼(𝐺’) may is the estimate of the true error rate for a classifier
How good is this estimator?
◼ Bias: How much does it deviate from the true value