Skip to content

Latest commit

 

History

History
17 lines (14 loc) · 1.51 KB

06_bias_variance.md

File metadata and controls

17 lines (14 loc) · 1.51 KB

Class 6 Pre-work: Bias-Variance Tradeoff

Read this excellent article, Understanding the Bias-Variance Tradeoff, and be prepared to discuss it in class on Monday.

Note: You can ignore sections 4.2 and 4.3.

Here are some questions to think about while you read:

  • In the Party Registration example, what are the features? What is the response? Is this a regression or classification problem?
  • Conceptually, how is KNN being applied to this problem to make a prediction?
  • How do the four visualizations in section 3 relate to one another? Change the value of K using the slider, and make sure you understand what changed in the visualizations (and why it changed).
  • In figures 4 and 5, what do the lighter colors versus the darker colors mean? How is the darkness calculated?
  • What does the black line in figure 5 represent? What predictions would an ideal machine learning model make, with respect to this line?
  • Choose a very small value of K, and click the button "Generate New Training Data" a number of times. Do you "see" low variance or high variance? Do you "see" low bias or high bias?
  • Repeat this with a very large value of K. Do you "see" low variance or high variance? Do you "see" low bias or high bias?
  • Try using other values of K. What value of K do you think is "best"? How do you define "best"?
  • Why should we care about variance at all? Shouldn't we just minimize bias and ignore variance?
  • Does a high value for K cause "overfitting" or "underfitting"?