Unsupervised Machine Learning

582638
5
Algoritmit ja koneoppiminen
Syventävät opinnot
Unsupervised learning is one of the main streams of machine learning, and closely related to multivariate statistics and data mining. This course describes some of the main methods in unsupervised learning, such as principal and independent component analysis, clustering, and nonlinear dimension reduction methods. In recent years, machine learning has become heavily dependent on statistical theory which is why this course is somewhere on the borderline between statistics and computer science. Emphasis is put both on the statistical/probabilistic formulation of the methods as well as on their computational implementation. The course is intended to CS students in the algorithms and machine learning specialisation, to statistics students, and to mathematics students in the statistical machine learning specialisation.

Koe

27.02.2013 09.00 B123
Vuosi Lukukausi Päivämäärä Periodi Kieli Vastuuhenkilö
2013 kevät 15.01-22.02. 3-3 Englanti Aapo Hyvärinen

Luennot

Aika Huone Luennoija Päivämäärä
Ti 14-16 C222 Aapo Hyvärinen 15.01.2013-22.02.2013
To 14-16 C222 Aapo Hyvärinen 15.01.2013-22.02.2013
Pe 14-16 C222 Aapo Hyvärinen 15.01.2013-22.02.2013

Information for international students

The course will be completely in English.

Yleistä

 

Please note that the course is in Period III this year.

 

Target audience

Master's students in computer science (specialization in algorithms & machine learning, or bioinformatics), applied mathematics (specialization statistical machine learning or e.g. stochastics), or statistics.

Description

Unsupervised learning is one of the main streams of machine learning, and closely related to exploratory data analysis and data mining. This course describes some of the main methods in unsupervised learning.

In recent years, machine learning has become heavily dependent on statistical theory which is why this course is somewhere on the borderline between statistics and computer science. Emphasis is put both on the statistical formulation of the methods as well as on their computational implementation. The goal is not only to introduce the methods on a theoretical level but also to show how they can be implemented in scientific computing environments such as Matlab or R.

Computer projects are an important part of the course, but they are given separate credits, see Projects for Unsupervised Machine Learning. The projects will be given in the exercice session marked below in the schedule.

One of the weekly sessions (usually Friday) will be an exercice session, the timetable will be as follows: 

Tue 15 Jan Lecture * Thu 17 Jan Lecture * Fri 18 Jan

Exercices

Intro to computer assignments

Tue 22 Jan Lecture * Thu 24 Jan Lecture * Fri 25 Jan Exercices
Tue 29 Jan Lecture * Thu 31 Jan Lecture * Fri 1 Feb Exercices
Tue 5 Feb Lecture * Thu 7 Feb Lecture * Fri 8 Feb Exercices
Tue 12 Feb Lecture * Thu 15 Feb Lecture * Fri 15 Feb Exercices
Tue 19 Feb Lecture * Thu 21 Feb Lecture * Fri 22 Feb Exercices

The exercices will be taught by Jouni Puuronen and the computer projects by Jukka-Pekka Kauppi.

Prerequisites

  • Statistics majors: Bachelor's degree recommended.
  • Mathematics majors: Bachelor's degree recommended. It should include basic courses in calculus (including vector calculus), linear algebra I&II, introduction to probability, introduction to statistical inference.
  • Computer science majors: Bachelor's degree recommended. It should include the following mathematics courses: calculus (including vector calculus), linear algebra I&II, introduction to probability. You should also have done "Introduction to machine learning" in Period II.

Contents:

  • Introduction. Supervised vs. unsupervised learning. Applications of unsupervised learning. Probabilistic formulation. Review of some basic mathematics (linear algebra, probability)
  • Numerical optimization. Gradient method, Newton's method, stochastic gradient, projected gradient methods
  • Principal component analysis and factor analysis. Formulation as minimization of reconstruction error or maximization of component variance. Computation using covariance matrix and its eigen-value decomposition. Factor analysis and interpretation of PCA as estimation of gaussian generative model. Factor rotations.
  • Independent component analysis. Problem of blind source separation, why non-gaussianity is needed for identifiability. Correlation vs. independence. ICA as maximization of non-gaussianity, measurement of non-Gaussianity by cumulants. Likelihood of the model and maximum likelihood estimation. Implementation by gradient methods and FastICA. Applications of component analysis.
  • Clustering. K-means algorithm.Gaussian mixture model: Maximization of likelihood, EM algorithm.
  • Nonlinear dimension reduction. Non-metric multi-dimensional scaling and related methods: kernel PCA, Laplacian eigenmaps, IsoMap. Kohonen's self-organizing map

Kurssin suorittaminen

There will be a single exam at the end of the course. Check the exact timetable and place on the CS dept exam page.

Active participation in the exercise sessions will give you points for the exam. See here for details.

 

Kirjallisuus ja materiaali

Here are the complete lecture notes for this year's course. Just to keep search engines away, you need the login uml and password uml. There is no book for the course.

Here are the exercices to be considered in the Friday sessions.