Schedule

  1. Introduction slides; assignment 01

    This lectures introduces machine learning, its applications, and the kind of problems it can be applied. Likewise, it presents related concepts such as supervised and unsupervised learning, and generalization.

  2. Linear and logistic regression lecture notes; lab 01: introduction, lab 02: regression; assignment 02

    This lecture introduces parametric approaches to supervised learning and linear models. Linear regressors are expressed as maximum likelihood estimation problem. The discussed concepts include: (a) parametric methods, (b) maximum likelihood estimates, (c) linear regression, and (d) logistic regression.

    References:

  3. Dimension reduction slides; lab 03: dimension reduction

    This lectures discusses how to tackle high-dimensional learning problems, and how to reduce dimension through the principal component analysis (PCA) method.

    References:

  4. Model evaluation and selection slides; lab 04: model assessment

    This lecture discusses model assessment of supervised machine learning. The discussed topics include: (a) training and test sets, (b) cross-validation, (c) bootstrap, (d) metrics of model complexity, and (e) metrics of performance for classification and regression.

  5. Regularized linear regression and nearest-neighbors methods slideslab 05: kNN

    This lecture introduces the concept of regularization as a means to controlling the complexity of the hypothesis space, and apply it to linear models. Furthermore, non-parametric methods are illustrated with the nearest-neighbors approaches. The discussed topics are: lasso, ridge regression, structured regularization, non-parametric learning, and k-nearest neighbors.

    References:

  6. Tree-based methods slides; lab 06.1: decision trees; lab 06.2: tree-based methods

    This lecture discusses decision tree approaches and shows how to combine simple classifiers to yield state-of-the-art predictors.

    References:

  7. Support vector machines

    This lecture introduces support-vector machine from its principles in the case of linearly separable data and shows how positive-definite kernels can be used to extend the approach to non-linear separating functions.

    References:

  8. Clustering

    This lecture introduces clustering, the common unsupervised learning problem. Its concepts are illustrated through hierarchical clustering, k-means, and DBSCAN.

  9. Practical Exam Jupyter Notebook