Course
Lab
Starting Unit 6 (Regularization/Penalized models) at end of class today; All set!
Unit 7 is Mid-term Exam unit
Minimal EDA with new data sets (and the last application assignment)
Most common scenario requires both model selection and final model evaluation
All methods (other than nested) use a single test set
initial_split()
for all but validation set approach (now use initial_validation_split()
for this last approach)Describe how each of these methods produce training and validation sets
When might you just want to know the best model configuration but NOT care about how it performs? (though I think you might always want to know!)
How would you modify the resampling procedure?
In this instance, we might call the held-out sets validation sets
when using that terminology
When might you have only one model configuration? (is this ever really true?)
How would you modify the resampling procedure?
In this instance, we might call the held-out sets test sets
when using that terminology
Describe these two properties of estimates in general
Describe bias and variance associated with developing models (estimates of the DGP)
Describe bias and variance of our performance estimates
Our resampling methods yield an ESTIMATE of (held-out/out of sample) performance metric
Lets consider the bias and variable of this ESTIMATE of a performance metric for each METHOD
Our METHODS
ALL resampling methods yield performance estimates with some degree of bias when used to evaluate the performance of a model we will implement trained on ALL the data
WHY?
What are the implications if you use that same performance estimate to select the best model configuration AND to evaluate that best model configuration (i.e., estimate its performance in new data)?
Can this source of bias be eliminated?
When used?
Why needed?
Describe how to do it using held-in/held-out terminology in addition to train/val/test
Discuss the bias and variance of the performance estimate used to select the best configuration. Consider implications of:
Discuss the bias and variance of the performance estimate used to evaluate that best configuration. Consider implications of: