PENN CIS 6250, FALL 2023: THEORY OF MACHINE LEARNING (original) (raw)
PENN CIS 6250, FALL 2023: THEORY OF MACHINE LEARNING
Prof. Michael Kearns
mkearns@cis.upenn.edu
Time: Tuesdays and Thursdays 10:15-11:45
Location: MacNeil 286-7
Attendance at lectures is a course requirement.
MK Office Hours: Right after class on Thursdays or by appointment. Weather permitting, I'll hold OHs in the outdoor seating area near MacNeil, otherwise in one of my offices. Please let me know in advance if you intend to come to OHs.
Teaching Assistants:
Brian Lee:wblee@wharton.upenn.edu
Office Hours: Tuesdays 12-1:30 in the Levine 4th floor bump space.
Keshav Ramji:keshavr@seas.upenn.edu
Office Hours: Wednesdays 5-6:30 in 612 Levine.
URL for this page:
www.cis.upenn.edu/~mkearns/teaching/CIS625
Previous incarnations of this course:
www.cis.upenn.edu/~mkearns/teaching/CIS625/cis6250-22.html
www.cis.upenn.edu/~mkearns/teaching/CIS625/cis625-21.html
www.cis.upenn.edu/~mkearns/teaching/COLT/colt18.html
www.cis.upenn.edu/~mkearns/teaching/COLT/colt17.html
www.cis.upenn.edu/~mkearns/teaching/COLT/colt16.html
www.cis.upenn.edu/~mkearns/teaching/COLT/colt15.html (with Grigory Yaroslavtsev)
www.cis.upenn.edu/~mkearns/teaching/COLT/colt12.html (with Jake Abernethy)
www.cis.upenn.edu/~mkearns/teaching/COLT/colt08.html (with Koby Crammer)
COURSE DESCRIPTION
This course is an introduction to the theory of machine learning, and provides mathematical, algorithmic, complexity-theoretic and probabilistic/statistical foundations to modern machine learning and related topics.
The first part of the course will follow portions ofAn Introduction to Computational Learning Theory,by M. Kearns and U. Vazirani (MIT Press). We will cover perhaps 6 or 7 of the chapters in K&V over (approximately) the first half of the course, often supplementing with additional readings and materials. I will provide electronic versions of the relevant chapters of K&V.
The second portion of the course will focus on a number of models and topics in learning theory and related areas not covered in K&V.
The course will give a broad overview of the kinds of theoretical problems and techniques typically studied and used in machine learning, and provide a basic arsenal of powerful mathematical tools for analyzing machine learning problems.
Topics likely to be covered include:
- Basics of the Probably Approximately Correct (PAC) Learning Model
- Occam's Razor, Compression and Learning
- Concentration, Uniform Convergence and the Vapnik-Chervonenkis Dimension
- Boosting and Related Techniques
- Learning in the Presence of Noise and Statistical Query Learning
- Learning and Cryptography
- Query Models
- Online and No-Regret Learning Models
- Agnostic Learning
- Game Theory and Machine Learning
- Learning and Differential Privacy
- Fairness in Machine Learning
- Theory of Reinforcement Learning