Slides from my intro to Bayesian regression talk
Back in April, I gave a guest lecture on Bayesian regression for the psychology department’s graduate statistics class. This is the same course where I first learned regression—and where I first started using R for statistics instead of for data cleaning. It was fun drawing on my experience in that course and tailoring the materials for the level of training.
Here are the materials:
Observations (training data)
As I did with my last Bayes talk, I’m going to note some questions from the audience, so I don’t forget what kinds of questions people have when they are introduced to Bayesian statistics.
One theme was frequentist baggage 👜. One person asked about Type I and Type II error rates. I did not have a satisfactory (that is, rehearsed) answer ready for this question. I think I said something about how those terms are based on a frequentist, repeated-sampling paradigm, whereas a Bayesian approach worries about different sorts of errors. (Statistical power is still important, of course, for both approaches.) Next time, I should study up on the frequentist properties of Bayesian models, so I can field these questions better.
- Another bit of frequentist baggage 👜. I mentioned that with a posterior predictive distribution, we can put an uncertainty interval on any statistic we can calculate, and this point brought up the question of multiple comparisons. These are a bad thing in classical statistics. But for Bayes, there is only one model, and the multiple comparisons are really only the implications of one model.
- Someone else said that they had heard that Bayesian models can provide evidence for a null effect—how does that work? I briefly described the ROPE approach, ignoring the existence of Bayes factors entirely.
For future iterations of this tutorial, I should have a worked example, maybe a blog post, on each of these issues.
It’s kind of amusing now that I think about it. A big part of my enthusiasm for Bayesian statistics is that I find it much more intuitive than frequentist statistics. Yes! I thought to myself. I never have to worry about what the hell a confidence interval is ever again! Well, actually—no. I need to know this stuff even more thoroughly than ever if I am going to talk fluently about what makes Bayes different. ¯\_(ツ)_/¯
Last knitted on 2022-05-27. Source code on GitHub.1
.session_info #> ─ Session info ─────────────────────────────────────────────────────────────── #> setting value #> version R version 4.2.0 (2022-04-22 ucrt) #> os Windows 10 x64 (build 22000) #> system x86_64, mingw32 #> ui RTerm #> language (EN) #> collate English_United States.utf8 #> ctype English_United States.utf8 #> tz America/Chicago #> date 2022-05-27 #> pandoc NA #> #> ─ Packages ─────────────────────────────────────────────────────────────────── #> package * version date (UTC) lib source #> assertthat 0.2.1 2019-03-21  CRAN (R 4.2.0) #> cli 3.3.0 2022-04-25  CRAN (R 4.2.0) #> crayon 1.5.1 2022-03-26  CRAN (R 4.2.0) #> emo 0.0.0.9000 2022-05-25  Github (hadley/emo@3f03b11) #> evaluate 0.15 2022-02-18  CRAN (R 4.2.0) #> generics 0.1.2 2022-01-31  CRAN (R 4.2.0) #> git2r 0.30.1 2022-03-16  CRAN (R 4.2.0) #> glue 1.6.2 2022-02-24  CRAN (R 4.2.0) #> here 1.0.1 2020-12-13  CRAN (R 4.2.0) #> knitr * 1.39 2022-04-26  CRAN (R 4.2.0) #> lubridate 1.8.0 2021-10-07  CRAN (R 4.2.0) #> magrittr 2.0.3 2022-03-30  CRAN (R 4.2.0) #> purrr 0.3.4 2020-04-17  CRAN (R 4.2.0) #> ragg 1.2.2 2022-02-21  CRAN (R 4.2.0) #> rlang 1.0.2 2022-03-04  CRAN (R 4.2.0) #> rprojroot 2.0.3 2022-04-02  CRAN (R 4.2.0) #> rstudioapi 0.13 2020-11-12  CRAN (R 4.2.0) #> sessioninfo 1.2.2 2021-12-06  CRAN (R 4.2.0) #> stringi 1.7.6 2021-11-29  CRAN (R 4.2.0) #> stringr 1.4.0 2019-02-10  CRAN (R 4.2.0) #> systemfonts 1.0.4 2022-02-11  CRAN (R 4.2.0) #> textshaping 0.3.6 2021-10-13  CRAN (R 4.2.0) #> xfun 0.31 2022-05-10  CRAN (R 4.2.0) #> #>  C:/Users/Tristan/AppData/Local/R/win-library/4.2 #>  C:/Program Files/R/R-4.2.0/library #> #> ──────────────────────────────────────────────────────────────────────────────
Leave a comment