EE 364A: Convex Optimization
This is an archive of materials used for CS 228T, taught at Stanford in 2011 under Daphne Koller.
Course description
Machine learning is a field at the intersection of computer science and statistics that aims to develop computational systems that learn from data and improve with experience. Though its origins lie in the field of artificial intelligence, modern machine learning has transformed a huge variety of areas, such as biology, medicine, e-commerce, retail, marketing, operations, logistics, politics, journalism, and, of course, finance.
This course provides a general introduction to machine learning with a view towards applications in finance. The goal is to provide both a solid grounding in the foundations of machine learning as well as a conceptual map of the field and its relation to areas like statistics and optimization. The focus is on mathematical and conceptual understanding; the course will occasionally touch on implementation issues and financial examples, but will not emphasize either aspect in coursework.
Topics include linear regression, logistic regression, exponential families, generalized linear models, generative models, support vector machines, loss functions and regularization, sparsity, Bayesian methods, model selection, the EM algorithm, clustering, principal components analysis, and convex optimization and optimization algorithms.
Prerequisites
The course requires background in linear algebra, probability, and optimization at the level of MATH 2940, ORIE 5500, and ORIE 5300.
Syllabus
- Introduction
- Convex optimization
- Supervised learning
- Linear regression
- Logistic regression
- Exponential families and generalized linear models
- Generative models for classification
- Support vector machines, duality, and kernelization
- Model selection, regularization, and Bayesian methods
- Optimization algorithms
- Unsupervised learning
The syllabus may be adjusted through the course of the semester.
Homework
- Problem Set 1, due February 13. Solutions.
- Problem Set 2, due February 27. Solutions.
- Problem Set 3, due March 13. Solutions.
- Problem Set 4, due March 27. Solutions.
- Problem Set 5, due April 26. Solutions.
- Problem Set 6, due May 8. Solutions.
Quizzes
- XXX
Readings
These readings will be posted intermittently through the semester and are entirely optional. Their goal is to give some exposure to the history, culture, and debates of machine learning, statistics, and data science, and to give additional perspective. Some are just included for historical interest and are not intended to be read cover to cover.
- A. Halevy, P. Norvig, and F. Pereira. The unreasonable effectiveness of data. IEEE Intelligent Systems, 2009.
- D. Mumford. The dawning of the age of stochasticity. From Mathematics towards the Third Millennium, 1999.
- J. Gleick. Breakthrough in problem solving. The New York Times, 1984.
- S. Stigler. Gauss and the invention of least squares. Annals of Statistics, 1981.
- L. Breiman. Statistical modeling: the two cultures. Statistical Science, 2001.
- T. Bayes. An essay towards solving a problem in the doctrine of chances. Philosophical Transactions of the Royal Society, 1764.
- B. Efron. Controversies in the foundation of statistics. American Mathematical Monthly, 1978.
- D. Freedman. Some issues in the foundation of statistics. Foundations of Science, 1995.
- D. Donoho. 50 years of data science. From Tukey Centennial Workshop, Princeton, NJ, 2015.
- Z. Tufekci. YouTube, the great radicalizer. The New York Times, 2018.