Machine Learning: The elegant way to extract information from data
Drs Jens Kauffmann, Veselina Kalinova, Dario Colombo and Tobias Albertson (MPIfR)
The human brain is able to recognize patterns in the data due to the millions of years of evolution. Today, we try to use and further develop computer algorithms that can reproduce this ability to distinguish patterns in large statistical samples and images, to find correlations between events, to classify objects based on their similar and different features in an objective and automatic way. Machine Learning is a sub-field in computer science that explores pattern recognition in the data analysis and characterize already known trends in sets of observations. This course focuses on the basic most applied up-to-date techniques in the machine learning that can be used in any branch of science and industry. We start with a broad introduction about the topic, followed by detailed discussion about the theory and application of the different types of the machine learning — supervised and unsupervised learning. In the field of supervised learning, for example, we will discuss regression, support vector machines, and neural networks, while the field of unsupervised learning we will cover clustering techniques, principal component analysis, and dimensionality reduction. The final goal of the course is to give the hands–on knowledge needed to pick and apply machine learning tools in Python. Discussions of the underlying mathematical principles will illustrate the inner workings of these tools.
Topics to be covered
Introduction — what is machine learning?
Supervised Learning — regression, support vector machines, neural networks
Unsupervised Learning — clustering, principal component analysis, dimensionality reduction
Tools — Monte Carlo Markov chains, Bayesian inference
Lecture 1: February 13 at 10:00 in 0.02: Introduction to Machine Learning: data mining, pattern recognition, dimensionality reduction, validation, objectivism | .mov .mp4
Lecture 2: February 14 at 10:00 in 0.02: Principal Component Analysis (PCA): eigenvectors, eigenvalues, variance, covariance, eigenfaces, application to astronomy | .mov .mp4
Lecture 3: February 15 at 10:00 in 0.02: Clustering I: distance definitions, hierarchical clustering, k-means | .mov .mp4
Lecture 4: February 16 at 10:00 in 0.02: Clustering II: graph theory, spectral clustering, DBSCAN, expectation maximization via gaussian mixture models, affinity propagation | .mov .mp4
Lecture 5: February 20 at 10:00 in 0.02: Regression: chi-squared fitting, Markov Chain Monte Carlo (MCMC), Bayesian inference | .mov .mp4
Lecture 6: February 21 at 10:00 in 0.02: Classification I: support vector machines, perceptron, artificial neural networks | .mov .mp4
Lecture 7: February 22 at 10:00 in 0.02: Classification II: decision trees, naive Bayes, genetic algorithm | .mov .mp4
Lecture 8: February 23 at 10:00 in 0.02: Summary: review of methods, validation,connections to astronomy, resources for further learning | .mov .mp4