Course Outline

list High School / Advanced Statistics and Data Science I (ABC)

Book
  • High School / Advanced Statistics and Data Science I (ABC)
  • High School / Statistics and Data Science I (AB)
  • High School / Statistics and Data Science II (XCD)
  • High School / Algebra + Data Science (G)
  • College / Introductory Statistics with R (ABC)
  • College / Advanced Statistics with R (ABCD)
  • College / Accelerated Statistics with R (XCD)
  • CKHub: Jupyter made easy

13.3 What You have Learned about Evaluating Models

Wow, look at what you have accomplished! You can come up with questions and hypotheses, fit models, and use fancy statistics like PRE and F in order to quantify how well a model does at explaining variation. But wait—you can do even more!

Our model is the best-fitting model for our data. But what we really want to know is this: Is there really a relationship between income and hate crimes in the Data Generating Process? After all, in any given year, month, or even week, the rate of hate crimes will change and incomes are constantly changing as well. Is this relationship true in the DGP? Or was the relationship in our sample possibly just caused by random chance?

A scatter plot of the distribution of avg_hatecrimes_per_100k_fbi by median_house_inc in hate_crimes overlaid with a best-fitting regression line.

Let’s consider our relatively better model, the income model of hate crimes. Is it really better than the empty model? Is our PRE of .10 significantly higher than we would expect if there were no relationship between income and hate crimes in the DGP?

You now know how to evaluate the income model compared to the empty model. More specifically, you know how to ask whether the relationship we observed between income and hate crimes is truly present in the DGP, or whether it could simply have come about by chance, a result of random sampling variation. It turns out that everything you need to know is right in the supernova table!

Analysis of Variance Table (Type III SS)
Model: avg_hatecrimes_per_100k_fbi ~ median_house_inc

                             SS df     MS     F    PRE     p
----- --------------- | ------- -- ------ ----- ------ -----
Model (error reduced) |  14.584  1 14.584 5.409 0.1013 .0243
Error (from model)    | 129.409 48  2.696
----- --------------- | ------- -- ------ ----- ------ -----
Total (empty model)   | 143.993 49  2.939

You now know how to imagine, and instantiate in R, a world where there is no relationship between income and hate crimes! You understand the concept of sampling distribution, and you know something about what supernova() was doing in order to calculate a p of .02.

But you learned how to do much more than just read and interpret the p-value in the supernova table. You also learned how to use different techniques for creating sampling distributions—techniques such as bootstrapping, randomization, and simulation.

Let’s use randomization to replicate what we found in the supernova table. To start, let’s create a sampling distribution with 1,000 randomized PREs. We’ve written code in the code window below to fit the linear model to our sample, and save the sample PRE as sample_PRE.

Add code to get 1,000 randomized samples (using shuffle()), and save their PREs in a data frame. Then, run a tally() to find the likelihood of getting a PRE greater than our sample statistic just by chance (it might be helpful to get the format in “proportions”). We have added code that will plot the sampling distribution in a histogram and fill the PREs greater than our sample PRE in a different color.

require(coursekata) hate_crimes <- hate_crimes %>% filter(complete.cases(avg_hatecrimes_per_100k_fbi, median_house_inc)) set.seed(41) # We've saved the sample PRE sample_PRE <- pre(avg_hatecrimes_per_100k_fbi ~ median_house_inc, data = hate_crimes) # Create a sampling distribution of 1000 randomized PREs SDoPRE <- # Run a tally to calculate p-value tally() # This code will depict the sampling distribution in a histogram gf_histogram(~pre, data = SDoPRE, fill = ~pre > sample_PRE) # We've saved the sample PRE sample_PRE <- pre(avg_hatecrimes_per_100k_fbi ~ median_house_inc, data = hate_crimes) # Create a sampling distribution of 1000 randomized PREs SDoPRE <- do(1000) * pre(shuffle(avg_hatecrimes_per_100k_fbi) ~ median_house_inc, data = hate_crimes) # NOTE: Best solution is: # Run a tally to calculate p-value # tally(~pre > sample_PRE, data = SDoPRE, format = "proportion") # This code will depict the sampling distribution in a histogram # gf_histogram(~pre, data = SDoPRE, fill = ~pre > sample_PRE) # This longer solution is there for scoring purposes and can be ignored. # Run a tally to calculate p-value tally(~pre > sample_PRE, data = SDoPRE, format = "proportion") # This code will depict the sampling distribution in a histogram gf_histogram(~pre, data = SDoPRE, fill = ~pre > sample_PRE) eq_fun <- function(x, y) (formula(x) == ~PRE > sample_PRE) ex() %>% { check_function(., "do") check_function(., "pre", index = 2) } ex() %>% check_or( check_function(., "gf_histogram") %>% { check_arg(., "data") %>% check_equal() check_arg(., "object") %>% check_equal() check_arg(., "fill") %>% check_equal(eval = FALSE, eq_fun = eq_fun) }, override_solution(., "gf_histogram(~pre, data = SDoPRE, fill = ~pre > sample_PRE)") %>% check_function("gf_histogram") %>% check_arg("fill") %>% check_equal(eval = FALSE, eq_fun = function (x,y) (formula(x) == ~pre > sample_PRE)) ) ex() %>% check_or( check_function(., "tally") %>% { check_arg(., "data") %>% check_equal() check_arg(., "format") %>% check_equal() check_arg(., "x") %>% check_equal(eval = FALSE, eq_fun = eq_fun) }, override_solution(., 'tally(~pre > sample_PRE, data = SDoPRE, format = "proportion")') %>% check_function("tally") %>% check_arg("x") %>% check_equal(eval = FALSE, eq_fun = function(x,y) (formula(x) == ~pre > sample_PRE)) )

A histogram of the distribution of PRE in the SDoPRE data frame. The area of PREs that is greater than sample_PRE is 0.023 and filled in blue. The rest of the area is 0.977 and filled in red.

pre > sample_PRE
 TRUE FALSE
0.023 0.977

You also know how to use randomization to generate a sampling distribution of F ratios, just like you did for PRE. Use the code window below to create a randomized sampling distribution of 1,000 Fs, and then calculate the probability that the F we observed in our sample could have been generated if the empty model is true.

require(coursekata) hate_crimes <- hate_crimes %>% filter(complete.cases(avg_hatecrimes_per_100k_fbi, median_house_inc)) set.seed(41) # We've saved the sample F sample_f <- f(avg_hatecrimes_per_100k_fbi ~ median_house_inc, data = hate_crimes) # Create sampling distribution of randomized Fs sdof <- # Run a tally to calculate p-value tally() # This code will depict the sampling distribution in a histogram gf_histogram(~f, data = sdof, fill = ~f > sample_f) # We've saved the sample F sample_f <- f(avg_hatecrimes_per_100k_fbi ~ median_house_inc, data = hate_crimes) # Create sampling distribution of randomized Fs sdof <- do(10) * f(shuffle(avg_hatecrimes_per_100k_fbi) ~ median_house_inc, data = hate_crimes) # NOTE: Best solution is: # Run a tally to calculate p-value # tally(~f > sample_f, data = sdof, format = "proportion") # This code will depict the sampling distribution in a histogram # gf_histogram(~f, data = sdof, fill = ~f > sample_f) # This longer solution is there for scoring purposes and can be ignored. # Run a tally to calculate p-value tally(~f > f(avg_hatecrimes_per_100k_fbi ~ median_house_inc, data = hate_crimes), data = sdof, format = "proportion") # This code will depict the sampling distribution in a histogram gf_histogram(~f, data = sdof, fill = ~f > f(avg_hatecrimes_per_100k_fbi ~ median_house_inc, data = hate_crimes)) eq_fun <- function(x, y) (formula(x) == ~f > f(avg_hatecrimes_per_100k_fbi ~ median_house_inc, data = hate_crimes)) ex() %>% { check_function(., "do") check_function(., "f", index = 2) } ex() %>% check_or( check_function(., "gf_histogram") %>% { check_arg(., "data") %>% check_equal() check_arg(., "object") %>% check_equal() check_arg(., "fill") %>% check_equal(eval = FALSE, eq_fun = eq_fun) }, override_solution(., "gf_histogram(~f, data = sdof, fill = ~f > sample_f)") %>% check_function("gf_histogram") %>% check_arg("fill") %>% check_equal(eval = FALSE, eq_fun = function(x,y) (formula(x) == ~f > sample_f)) ) ex() %>% check_or( check_function(., "tally") %>% { check_arg(., "data") %>% check_equal() check_arg(., "format") %>% check_equal() check_arg(., "x") %>% check_equal(eval = FALSE, eq_fun = eq_fun) }, override_solution(., 'tally(~f > sample_f, data = sdof, format = "proportion")') %>% check_function("tally") %>% check_arg("x") %>% check_equal(eval = FALSE, eq_fun = function(x,y) (formula(x) == ~f > sample_f)) )

A histogram of the distribution of f in the sdof data frame. The area of Fs that is greater than sample_f is 0.023 and filled in blue. The rest of the area is 0.977 and filled in red.

f > sample_f
 TRUE FALSE
0.023 0.977

Pretty amazing. We started with a mere sample, a brief glimpse of what the Data Generating Process of hate crimes might be. But you showed that there is only a small likelihood (.02) of generating our PRE and F ratio if there is no relationship between income and hate crimes in the world. It’s an amazing thing that you can now take a sample and have sophisticated tools to say something about the population!

But we always want you to stay humble about your statistical know-how. Remember, there are always limits to what you can learn from statistics.

Closing Words

At the beginning of this course we told you that exploring variation, modeling variation, and evaluating models organizes pretty much everything you need to know about statistics. You have truly come a long way since you first heard these vague ideas. You have just demonstrated how much you have grown your statistical skills and understanding! Look at you, asking important questions with data, making amazing plots with just a few keystrokes, fitting models, using fancy GLM notation, creating sampling distributions, and interpreting your analyses with humility. Thank you for all your efforts in grappling with hard concepts. We hope you continue to pursue real understanding in this topic and beyond!

Responses