Course Outline

list College / Advanced Statistics with R (ABCD)

Book College / Advanced Statistics with R (ABCD)
  • High School / Advanced Statistics and Data Science I (ABC)
  • High School / Statistics and Data Science I (AB)
  • High School / Statistics and Data Science II (XCD)
  • High School / Algebra + Data Science (G)
  • College / Introductory Statistics with R (ABC)
  • College / Advanced Statistics with R (ABCD)
  • College / Accelerated Statistics with R (XCD)
  • CKHub: Jupyter made easy

12.11 Confidence Interval for the Slope of a Regression Line

Let’s go back to the regression model we fit using FoodQuality to predict Tip. We can specify this model of the DGP like this:

Yi=β0+β1Xi+ϵi

Here is the lm() output for the best-fitting FoodQuality model.

Call:
lm(formula = Tip ~ FoodQuality, data = TipExperiment)

Coefficients:
 (Intercept)   FoodQuality  
     10.1076        0.3776  

Use the code window below to find the 95% confidence interval for the slope of this regression line.

                   2.5 %     97.5 %
(Intercept)  -9.29657877 29.4923793
FoodQuality   0.01546542  0.7400759

β1 represents the increment that is added to the predicted tip percent in the DGP for every additional point of food quality rating. The confidence interval of β1 represents the range of β1s from which our sample b1 is still likely (that is, not unlikely). β1s as low as 0.15 and as high as 0.74 can both reasonably produce our sample b1.

Now that we have tried confint(), try using the resample() function to bootstrap the 95% confidence interval for the slope of the regression line. See how your bootstrapped confidence interval compares to the results obtained by using confint().

Here is a histogram of the bootstrapped sampling distribution we created. Yours will be a little different, of course, because it is random.

A histogram of the sampling distribution of b1. It is normal in shape, centered near 0.38, and ranges from about -0.2 to 1.0. The tails of the distribution are shaded in teal, and the middle 95 percent of b1s are shaded in purple. The teal of the lower tail extends from about -0.2 to about 0.01, and the blue of the upper tail extends from about 0.75 to about 1.0. The range along the y-axis extends from zero to 80.

The center of the bootstrapped sampling distribution is approximately the same as the sample b1 of .38. This is what we would expect because bootstrapping assumes that the sample is representative of the DGP.

As explained previously, we can use the .025 cutoffs that separate the unlikely tails from the likely middle of the sampling distribution as a handy way to find the lower and upper bound of the 95% confidence interval. We can eyeball these cutoffs by looking at the histogram, or we can calculate them by arranging the bootstrapped sampling distribution to find the actual 26th and 975th b1s.

0.0198060804221204
0.732391298337459

To find the confidence interval, we sorted the randomly generated b1s from lowest to highest, and then used the 26th and 975th b1s as the lower and upper bounds of the confidence interval. Your results will be a little different from ours because resampling is random. We got a bootstrapped confidence interval of .02 to .73, which is close to what we got from confint() (.02 and .74).

The bootstrapped sampling distribution of slopes in this case is not exactly symmetrical; it is a bit skewed to the right. For this reason, the center of the confidence interval will not be exactly at the sample b1. This is in contrast to the mathematical approach that assumes that the sample b1 is exactly in the middle of a perfectly symmetrical t-distribution. This difference does not mean that bootstrapping is less accurate. It might be that there is something about the distributions of FoodQuality and Tip that results in this asymmetry.

The important thing we want to focus on for now is that all of these methods result in approximately the same results. These similarities show us what confidence intervals mean and what they can tell us. Later, in more advanced courses, you can take up the question of why the results differ across methods when they do.

Responses