统计学习导论_统计学习导论基于R应用 Slides

An Introduction to Statistical Learning with Applications in R

In January 2014, Stanford University professors Trevor Hastie and Rob Tibshirani (authors of the legendary Elements of Statistical Learning textbook) taught an online course based on their newest textbook, An Introduction to Statistical Learning with Applications in R (ISLR). I found it to be an excellent course in statistical learning (also known as “machine learning”), largely due to the high quality of both the textbook and the video lectures. And as an R user, it was extremely helpful that they included R code to demonstrate most of the techniques described in the book.

If you are new to machine learning (and even if you are not an R user), I highly recommend reading ISLR from cover-to-cover to gain both a theoretical and practical understanding of many important methods for regression and classification. It is available as a free PDF download from the authors’ website.

If you decide to attempt the exercises at the end of each chapter, there is a GitHub repository of solutions provided by students you can use to check your work.

As a supplement to the textbook, you may also want to watch the excellent course lecture videos (linked below), in which Dr. Hastie and Dr. Tibshirani discuss much of the material. In case you want to browse the lecture content, I’ve also linked to the PDF slides used in the videos.

Chapter 1: Introduction (slides, playlist)

  • Opening Remarks and Examples (18:18)
  • Supervised and Unsupervised Learning (12:12)

Chapter 2: Statistical Learning (slides, playlist)

  • Statistical Learning and Regression (11:41)
  • Curse of Dimensionality and Parametric Models (11:40)
  • Assessing Model Accuracy and Bias-Variance Trade-off (10:04)
  • Classification Problems and K-Nearest Neighbors (15:37)
  • Lab: Introduction to R (14:12)

Chapter 3: Linear Regression (slides, playlist)

  • Simple Linear Regression and Confidence Intervals (13:01)
  • Hypothesis Testing (8:24)
  • Multiple Linear Regression and Interpreting Regression Coefficients (15:38)
  • Model Selection and Qualitative Predictors (14:51)
  • Interactions and Nonlinearity (14:16)
  • Lab: Linear Regression (22:10)

Chapter 4: Classification (slides, playlist)

  • Introduction to Classification (10:25)
  • Logistic Regression and Maximum Likelihood (9:07)
  • Multivariate Logistic Regression and Confounding (9:53)
  • Case-Control Sampling and Multiclass Logistic Regression (7:28)
  • Linear Discriminant Analysis and Bayes Theorem (7:12)
  • Univariate Linear Discriminant Analysis (7:37)
  • Multivariate Linear Discriminant Analysis and ROC Curves (17:42)
  • Quadratic Discriminant Analysis and Naive Bayes (10:07)
  • Lab: Logistic Regression (10:14)
  • Lab: Linear Discriminant Analysis (8:22)
  • Lab: K-Nearest Neighbors (5:01)

Chapter 5: Resampling Methods (slides, playlist)

  • Estimating Prediction Error and Validation Set Approach (14:01)
  • K-fold Cross-Validation (13:33)
  • Cross-Validation: The Right and Wrong Ways (10:07)
  • The Bootstrap (11:29)
  • More on the Bootstrap (14:35)
  • Lab: Cross-Validation (11:21)
  • Lab: The Bootstrap (7:40)

Chapter 6: Linear Model Selection and Regularization (slides, playlist)

  • Linear Model Selection and Best Subset Selection (13:44)
  • Forward Stepwise Selection (12:26)
  • Backward Stepwise Selection (5:26)
  • Estimating Test Error Using Mallow’s Cp, AIC, BIC, Adjusted R-squared (14:06)
  • Estimating Test Error Using Cross-Validation (8:43)
  • Shrinkage Methods and Ridge Regression (12:37)
  • The Lasso (15:21)
  • Tuning Parameter Selection for Ridge Regression and Lasso (5:27)
  • Dimension Reduction (4:45)
  • Principal Components Regression and Partial Least Squares (15:48)
  • Lab: Best Subset Selection (10:36)
  • Lab: Forward Stepwise Selection and Model Selection Using Validation Set (10:32)
  • Lab: Model Selection Using Cross-Validation (5:32)
  • Lab: Ridge Regression and Lasso (16:34)

Chapter 7: Moving Beyond Linearity (slides, playlist)

  • Polynomial Regression and Step Functions (14:59)
  • Piecewise Polynomials and Splines (13:13)
  • Smoothing Splines (10:10)
  • Local Regression and Generalized Additive Models (10:45)
  • Lab: Polynomials (21:11)
  • Lab: Splines and Generalized Additive Models (12:15)

Chapter 8: Tree-Based Methods (slides, playlist)

  • Decision Trees (14:37)
  • Pruning a Decision Tree (11:45)
  • Classification Trees and Comparison with Linear Models (11:00)
  • Bootstrap Aggregation (Bagging) and Random Forests (13:45)
  • Boosting and Variable Importance (12:03)
  • Lab: Decision Trees (10:13)
  • Lab: Random Forests and Boosting (15:35)

Chapter 9: Support Vector Machines (slides, playlist)

  • Maximal Margin Classifier (11:35)
  • Support Vector Classifier (8:04)
  • Kernels and Support Vector Machines (15:04)
  • Example and Comparison with Logistic Regression (14:47)
  • Lab: Support Vector Machine for Classification (10:13)
  • Lab: Nonlinear Support Vector Machine (7:54)

Chapter 10: Unsupervised Learning (slides, playlist)

  • Unsupervised Learning and Principal Components Analysis (12:37)
  • Exploring Principal Components Analysis and Proportion of Variance Explained (17:39)
  • K-means Clustering (17:17)
  • Hierarchical Clustering (14:45)
  • Breast Cancer Example of Hierarchical Clustering (9:24)
  • Lab: Principal Components Analysis (6:28)
  • Lab: K-means Clustering (6:31)
  • Lab: Hierarchical Clustering (6:33)

Interviews (playlist)

  • Interview with John Chambers (10:20)
  • Interview with Bradley Efron (12:08)
  • Interview with Jerome Friedman (10:29)
  • Interviews with statistics graduate students (7:44)

a813fa52b2b6c8c86ba1ad8bf92b2200.png
Statistical learning refers to a set of tools for modeling and understanding complex datasets. It is a recently developed area in statistics, and blends with parallel developments in computer science, and in particular machine learning. The field encompasses many methods such as the lasso and sparse regression, classification and regression trees, and boosting and support vector machines. With the explosion of “Big Data” problems statistical learning has be- come a very hot field in many scientific areas as well as marketing, finance and other business disciplines. People with statistical learning skills are in high demand. One of the first books in this area — The Elements of Statistical Learn- ing (ESL) (Hastie, Tibshirani, and Friedman) — was published in 2001, with a second edition in 2009. ESL has become a popular text not only in statistics but also in related fields. One of the reasons for ESL’s popu- larity is its relatively accessible style. But ESL is intended for individuals with advanced training in the mathematical sciences. An Introduction to Statistical Learning (ISL) arose from the perceived need for a broader and less technical treatment of these topics. In this new book, we cover many of the same topics as ESL, but we concentrate more on the applications of the methods and less on the mathematical details. We have created labs illustrating how to implement each of the statistical learning methods using the popular statistical software package R . These labs provide the reader with valuable hands-on experience. This book is appropriate for advanced undergraduates or master’s stu- dents in Statistics or related quantitative fields, or for individuals in other disciplines who wish to use statistical learning tools to analyze their data. It can be used as a textbook for a course spanning one or two semesters. We would like to thank several readers for valuable comments on prelim- inary drafts of this book: Pallavi Basu, Alexandra Chouldechova, Patrick Danaher, Will Fithian, Luella Fu, Sam Gross, Max Grazier G’Sell, Court- ney Paulson, Xinghao Qiao, Elisa Sheng, Noah Simon, Kean Ming Tan, Xin Lu Tan. It’s tough to make predictions, especially about the future. -Yogi Berra
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值