Skip to content

A series of Python Jupyter notebooks that help you better understand "The Elements of Statistical Learning" book

Notifications You must be signed in to change notification settings

maxvv/The-Elements-of-Statistical-Learning-Python-Notebooks

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

"The Elements of Statistical Learning" Notebooks

Reproducing examples from the "The Elements of Statistical Learning" by Trevor Hastie, Robert Tibshirani and Jerome Friedman with Python and its popular libraries: numpy, math, scipy, sklearn, pandas, tensorflow, statsmodels, sympy, catboost, pyearth, mlxtend, cvxpy. Almost all plotting is done using matplotlib, sometimes using seaborn.

Examples

The documented Jupyter Notebooks are in the examples folder:

Classifying the points from a mixture of "gaussians" using linear regression, nearest-neighbor, logistic regression with natural cubic splines basis expansion, neural networks, support vector machines, flexible discriminant analysis over MARS regression, mixture discriminant analysis, k-Means clustering, Gaussian mixture model and random forests.

alt

Predicting prostate specific antigen using ordinary least squares, ridge/lasso regularized linear regression, principal components regression, partial least squares and best subset regression. Model parameters are selected by K-folds cross-validation.

alt

Understanding the risk factors using logistic regression, L1 regularized logistic regression, natural cubic splines basis expansion for nonlinearities, thin-plate spline for mutual dependency, local logistic regression, kernel density estimation and gaussian mixture models.

alt

Vowel speech recognition using regression of an indicator matrix, linear/quadratic/regularized/reduced-rank discriminant analysis and logistic regression.

alt

Comparing patterns of bone mineral density relative change for men and women using smoothing splines.

alt

Analysing Los Angeles pollution data using smoothing splines.

alt

Phonemes speech recognition using reduced flexibility logistic regression.

alt

Analysing radial velocity of galaxy NGC7531 using local regression in multidimentional space.

alt

Analysing the factors influencing ozone concentration using local regression and trellis plot.

alt

Detecting email spam using logistic regression, generalized additive logistic model, decision tree, multivariate adaptive regression splines, boosting and random forest.

alt

Analysing the factors influencing California houses prices using boosting over decision trees and partial dependance plots.

alt

Predicting shopping mall customers occupation, and hence identifying demographic variables that discriminate between different occupational categories using boosting and market basket analysis.

alt

Recognizing small hand-drawn digits using LeCun's Net-1 - Net-5 neural networks.

alt

Analysing of the number three variation in ZIP codes using principal component and archetypal analysis.

alt

Analysing microarray data using K-means clustring and hierarchical clustering.

alt

Analysing country dissimilarities using K-medoids clustering and multidimensional scaling.

alt

Analysing signature shapes using Procrustes transformation.

alt

Recognizing wave classes using linear, quadratic, flexible (over MARS regression), mixture discriminant analysis and decision trees.

alt

Analysing protein flow-cytometry data using graphical-lasso undirected graphical model for continuous variables.

alt

Analysing microarray data of 2308 genes and selecting the most significant genes for cancer classification using nearest shrunken centroids.

alt

Analysing microarray data of 16,063 genes gathered by Ramaswamy et al. (2001) and selecting the most significant genes for cancer classification using nearest shrunken centroids, L2-penalized discriminant analysis, support vector classifier, k-nearest neighbors, L2-penalized multinominal, L1-penalized multinominal and elastic-net penalized multinominal. It is a difficult classification problem with p>>N (only 144 training observations).

Solving a synthetic classification problem using Support Vector Machines and multivariate adaptive regression splines to show the influence of additional noise features.

Assessing the significance of 12,625 genes from microarray study of radiation sensitivity using Benjamini-Hochberg method and the significane analysis of microarrays (SAM) approach.

alt

About

A series of Python Jupyter notebooks that help you better understand "The Elements of Statistical Learning" book

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 100.0%