jan5 - Sta 414/2104 S...

Info iconThis preview shows pages 1–6. Sign up to view the full content.

View Full Document Right Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon

Info iconThis preview has intentionally blurred sections. Sign up to view the full version.

View Full DocumentRight Arrow Icon
This is the end of the preview. Sign up to access the rest of the document.

Unformatted text preview: Sta 414/2104 S http://www.utstat.utoronto.ca/reid/414S10.html STA414S/2104S: Statistical Methods for Data Mining and Machine Learning January - April, 2010 , Tuesday 12-2, Thursday 12-1, SS 2105 Course Information This course will consider topics in statistics that have played a role in the development of techniques for data mining and machine learning. We will cover linear methods for regression and classification, nonparametric regression and classification methods, generalized additive models, aspects of model inference and model selection, model averaging and tree based methods. Prerequisite : Either STA 302H (regression) or CSC 411H (machine learning). CSC108H was recently added: this is not urgent but you must be willing to use a statistical computing environment such as R or Matlab. Office Hours : Tuesdays, 34; Thursdays, 23; or by appointment. Textbook : Hastie, Tibshirani and Friedman. The Elements of Statistical Learning . SpringerVerlag. http://wwwstat.stanford.edu/~tibs/ElemStatLearn/index.html Course evaluation : Homework 1 due February 11: 20%, Homework 2 due March 4: 20%, Midterm exam, March 16: 20%, Final project due April 16: 40%. 1 / 20 Tentative Syllabus I Regression : linear, ridge, lasso, logistic, polynomial splines, smoothing splines, kernel methods, additive models, regression trees, projection pursuit, neural networks: Chapters 3, 5, 9, 11 I Classification : logistic regression, linear discriminant analysis, generalized additive models, kernel methods, naive Bayes, classification trees, support vector machines, neural networks, K-means, k-nearest neighbours, random forests: Chapters 4, 6, 9, 11, 12, 15 I Model Selection and Averaging : AIC, cross-validation, test error, training error, bootstrap aggregation: Chapter 7, 8.7 I Unsupervised learning : Kmeans clustering, k-nearest neighbours, hierarchical clustering: Chapter 14 2 / 20 Some references I Venables, W.N. and Ripley, B.D. (2002). Modern Applied Statistics with S (4th Ed.) . Springer-Verlag. Detailed reference for computing with R . I Maindonald, J. and Braun, J. (). Data Analysis and Graphics using R . Cambridge University Press. Gentler source for R information. I Hand, D., Mannila, H. and Smyth, P. (2001). Principals of Data Mining . MIT Press. Nice blend of computer science and statistical methods. I Ripley, B.D. (1996). Pattern Recognition and Neural Networks. Cambridge University Press. Excellent, but concise, discussion of many machine learning methods. 3 / 20 Some Courses I http://www.stat.columbia.edu/ madigan/DM08 Columbia (with links to other courses) I http://www-stat.stanford.edu/ tibs/stat315a.html Stanford I http://www.stat.cmu.edu/ larry/=sml2008/ Carnegie-Mellon 4 / 20 Data mining/machine learning I large data sets I high dimensional spaces I potentially little information on structure I computationally intensive I plots are essential, but require considerable pre-processing I emphasis on means and variances I emphasis on prediction...
View Full Document

Page1 / 20

jan5 - Sta 414/2104 S...

This preview shows document pages 1 - 6. Sign up to view the full document.

View Full Document Right Arrow Icon
Ask a homework question - tutors are online