Time and Place: 
Mon-Wed 5 - 6:20 PM in Peter 103 (Peterson Hall).

Instructor: 
Raef Bassily 
Email: rbassily at ucsd dot edu
Office Hrs: Thu 5-6 PM, Atkinson Hall 4111.

Course Schedule:
(Times and contents may slightly vary depending on the class progress) 

 Week 1: 3/ 28, 30 - Introduction and Course Overview
- Probability Tools and Concentration Inequalities
    Lecture notes (Part 1)
 Week 2: 4/ 4, 6 - Framework of Statistical Learning, Empirical Risk Minimization
- PAC Learning Model
 Week 3: 4/ 11, 13 - PAC Learning, Occam's Razor
- Agnostic PAC Learning and the Uniform Convergence Principle
    Lecture notes (Part 2) 
 Week 4: 4/ 18, 20 - Set Shattering: Intro to Vapnik-Chervonenkis (VC) dimension
- VC dimension: Examples, Discussion and Implications, and the Growth function
 Week 5: 4/ 25, 27 - VC dimension: Sauer's Lemma 
- The Fundamental Theorem of Learning (Characterization of Learnability via VC dimension).
    Lecture notes (Part 3)
- Boosting: Weak vs. Strong Learnability
 Week 6: 5/ 2, 4 - Boosting: Adaboost
    Lecture notes (Part 4)
Agnostic-PAC Learning in the Generalized Loss Model
 Week 7: 5/ 9, 11 - Midterm (on May 9th)
- Brief Intro to Convex Analysis: Convex, Lipschitz functions.
 
Week 8: 5/ 16, 18
- Learnability of Convex-Lipschitz-Bounded Problems
- Stochastic Gradient Descent:
    * Basic GD Algorithm and Convergence Guarantees. 
    * Projected Stochastic Gradient Descent.
- Learning via Stochastic Gradient Descent.
SGD Part I
SGD Part II
Weeks 9 & 10: 5/ 23, 25 & 6/ 1  - Regularization and Stability
    * Regularized Loss Minimization and Balancing Bias-Complexity
    * Regularization as a Stabilizer: Stable algorithms do not overfit.
    * Learning via RLM

Announcements: 


Course Overview: 

The course will be aiming mainly at explaining the main concepts underlying machine learning and the techniques that transform such concepts into practical algorithms. The main focus will be on the theoretical foundations of the subject, and the material covered will contain rigorous mathematical proofs and analyses. The class will cover wide array of topics starting from the basic models and concepts: PAC learning, uniform convergence, generalization, VC dimensions, and building on those to discuss more complex models and algorithmic techniques, e.g., Boosting, Convex Learning, Regularization, and Stochastic Gradient Descent.


Prerequisites:

Decent knowledge of probability and multivariate calculus is required. Students should be comfortable working with mathematical abstractions and proofs. Some previous exposure to machine learning is recommended. 


Homeworks:

Grading Policy: 

Discussion Forum: 

Please sign-up here to join the discussion forum of this class on Piazza.


Supplementary readings: