Course Information
Short Intro
The objective of this course is to provide a complete introduction to
machine learning, which discusses computer programs that automatically
improve their performance through experience (e.g., programs that learn
to recognize faces, recommend music and movies, and drive autonomous
robots).
Description
This course covers theoretical and practical algorithms for machine
learning from different perspectives: theoretical concepts include
induction bias, PAC learning framework, Bayesian learning methods, and
Occam’s razor; programming assignments include hands-on experiments with
various learning algorithms. This course is designed to provide a solid
foundation for students who wish to master modern artificial
intelligence techniques and to provide the necessary methods,
techniques, mathematics, and algorithms for those who wish to pursue
future research in machine learning.
The course will touch on the following topics:
- decision tree learning, linear models
- model selection and optimization, statistical learning methods
- neural networks, deep learning
- nonparametric models, ensemble learning
Keywords: machine learning, statistical learning,
deep learning
Course group
Previous offers: [2023 Spring]
Closely related: [Artificial
Intelligence], [Deep Learning], [Natural Language Processing]
Prerequisites
We expect you to have the following skills before taking this
course:
Required:
- Linear algebra (vector, matrix computation, Euclidean spaces).
- Differential calculus (Jacobian, Hessian, chain rule).
- Probabilities and statistics (common distributions, law of large
numbers, conditional probabilities, Bayes)
- Proficient in one programming language (Python recommended).
Recommended:
- Numerical optimization (notion of minima, gradient descent).
- Algorithm analysis (computational costs).
- Specialized knowledge (visual computing, robotics, speech and
language processing).
Teaching Plan
This course is organized into a 12-week session (4 hours per week).
The main contents are listed below:
- Introduction 导言,Preliminaries 预备知识
- Decision Trees 决策树
- Linear Models 线性模型
- Model Selection 模型选择
- (*)Nonparametric Models 非参数化模型
- (*)Ensemble Learning 集成学习
- (*)Statistical Learning Concepts 统计学习概念
- (*)Probability Models: Complete Data 概率模型:完备数据
- (*)Probability Models: Hidden Variables 概率模型:隐变量
- Feedforward Neural Networks 前馈神经网络,Back Propagation
反向传播
- Deep Learning Computation 深度学习计算
- Convolutional Neural Networks 卷积神经网络
- Modern Convolutional Neural Networks 现代卷积神经网络
- Recurrent Neural Networks 循环神经网络
- Modern Recurrent Neural Networks 现代循环神经网络
Tutorials are designed to consistent with lectures:
- Getting Started 实验配置及基础操作
- Decision Trees 决策树
- Linear Regression 线性回归
- Linear Classification 线性分类
- (*)Nonparametric Models 非参数化模型
- (*)Ensemble Learning 集成学习
- Feedforward Neural Networks 前馈神经网络
- Deep Learning Computation 深度学习计算
- Convolutional Neural Networks 卷积神经网络
- Recurrent Neural Networks 循环神经网络
Schedule
Tuesday, Thursday; YG09-404.
3 |
2024/03/11 |
[课程信息] [导言] |
机器学习范式、基本流程 |
|
2024/03/14 |
[导言] [预备知识] [安装配置] |
深度学习 |
4 |
2024/03/18 |
[决策树] |
决策树的表达力 |
|
2024/03/21 |
[决策树] |
ID3算法 |
5 |
2024/03/25 |
[线性回归] |
最小二乘法、多元微分 |
|
2024/03/25 |
[线性回归] |
梯度下降法 |
6 |
2024/04/01 |
[线性分类] |
感知机 |
7 |
2024/04/08 |
[线性分类] |
逻辑回归 |
|
2024/04/11 |
[前馈神经网络] |
反向传播算法 |
8 |
2024/04/15 |
[前馈神经网络] |
多层感知机 |
|
2024/04/18 |
[模型选择] |
模型选择 |
9 |
2024/04/22 |
[模型选择] |
欠拟合与过拟合 |
|
2024/04/25 |
[深度学习计算] |
数值稳定性 |
|
2024/04/28 |
[深度学习计算] |
参数管理 |
10 |
2024/04/29 |
[卷积神经网络] |
图像卷积 |
11 |
2024/05/06 |
[卷积神经网络] |
卷积设计 |
12 |
2024/05/13 |
[现代卷积神经网络] |
LeNet、GoogLeNet |
|
2024/05/16 |
[现代卷积神经网络] |
ResNet |
13 |
2024/05/20 |
[计算机视觉I] |
图像增广、微调、锚框 |
|
2024/05/23 |
[计算机视觉I] |
锚框 |
14 |
2024/05/27 |
[计算机视觉II] |
目标检测算法、语义分割、风格迁移 |
|
2024/05/30 |
[计算机视觉II] |
|
Several special topics:
Methodology
Problem-solving oriented, equal emphasis on lecture and practice.
以解决实际问题为导向,教学与实践并重。
Each lecture is roughly organized into 3 progressive units:
- Core Concepts 核心概念: provides elementary knowledge of the
topic
- Advanced Discussion 进阶讨论: provides in-depth understanding,
mathematical formulations
- Practical Skills 实践技巧: provides problem-solving skills through
hands-on programming training
Evaluation
- Attendance & participation: 10%
- Understanding of topics: 20%
- Practical exercises: 20%
- Final exam/project: 50%
Project
Textbook
Not mandatory but recommended:
- Christopher Bishop, Pattern Recognition and Machine
Learning, 2006, aka. PRML
- Bishop, C.M. and Bishop, H. Deep Learning: Foundations and
Concepts. Springer, 2023. aka. “New Bishop”
- Zhang et al., Dive into Deep Learning, 2023,
即《动手学深度学习》
- Hastie, T., Tibshirani, R., & Friedman, J. The Elements of
Statistical Learning: Data Mining, Inference, and Prediction (2nd
ed.). Stanford, 2009. 即《统计学习要素:
机器学习中的数据挖掘、推断和预测》
- Murphy, Probabilistic Machine Learning: An Introduction &
Advanced Topics, 2022
Resource
- 动手学深度学习
- Machine Learning with PyTorch and Scikit-Learn Book: GitHub, online courses
- PRML Book
- Deep Learning: Foundations and Concepts
- The Elements of Statistical Learning
- Probabilistic Machine Learning: An Introduction,
Advanced Topics
- Mathematics for Machine Learning