Faculty of Information Technology, BUT

Course details

Classification and Recognition

IKR Acad. year 2018/2019 Summer semester 5 credits

Current academic year

The tasks of classification and pattern recognition, basic schema of a classifier, data and evaluation of individual methods, statistical pattern recognition, feature extraction, multivariate Gaussian distribution,, maximum likelihood estimation, Gaussian Mixture Model (GMM), Expectation Maximization (EM) algorithm, linear classifiers, perceptron, Gaussian Linear Classifier, logistic regression, support vector machines (SVM), feed-forward neural networks, convolutional and recurrent neural networks, sequence classification, Hidden Markov Models (HMM). Applications of the methods to speech and image processing.

Guarantor

Deputy Guarantor

Language of instruction

Czech

Completion

Examination (written+oral)

Time span

26 hrs lectures, 13 hrs exercises, 13 hrs projects

Assessment points

60 exam, 15 half-term test, 25 projects

Department

Lecturer

Instructor

Subject specific learning outcomes and competences

The students will get acquainted with the problem of machine learning applied to pattern classification and recognition. They will learn how to apply basic methods in the fields of speech processing and computer graphics. They will understand the common aspects and differences of the particular methods and will be able to take advantage of the existing classifiers in real-situations.

Generic learning outcomes and competences

The students will get acquainted with python libraries focused on math, linear algebra and machine learning. They will also improve their math skills (probability theory, statistics, linear algebra) a programming skills. The students will learn to work in a team.

Learning objectives

To understand the foundations of machine learning with the focus on pattern classification and recognition. To learn how to apply basic algorithms and methods from this field to problems in speech and image recognition. To conceive basic principles of different generative an discriminative models for statistical pattern recognition. To get acquainted with the evaluation procedures.

Why is the course taught

Recent years witness a boom of machine learning or pattern recognition applications. More and more devices can be controlled using voice or gestures. Digital cameras automatically detect faces in the captured images in order to automatically focus or somehow react on it. Virtual agents in mobile devices can recognize speech and search for relevant answers to our queries. The quality of the current systems for automatic recognition of person's identity from voice recording or from face photo already significantly exceed the human abilities.

In this class, the students should learn how these technologies works. They will learn about the basic algorithms and models, which, using some training examples, automatically learn to recognize nontrivial patterns in audio recordings, images or other signals or input data.

Prerequisites

Prerequisite kwnowledge and skills

Basic knowledge of the standard math notation.

Study literature

  • Bishop, C. M.: Pattern Recognition, Springer Science + Business Media, LLC, 2006, ISBN 0-387-31073-8.
  • Hart, P. E., Stork, D. G.:Pattern Classification (2nd ed), John Wiley &Sons, 2000, ISBN: 978-0-471-05669-0

Fundamental literature

  • Bishop, C. M.: Pattern Recognition, Springer Science + Business Media, LLC, 2006, ISBN 0-387-31073-8.
  • Hart, P. E., Stork, D. G.:Pattern Classification (2nd ed), John Wiley &
    Sons, 2000, ISBN: 978-0-471-05669-0
  • Ian Goodfellow, Yoshua Bengio, and Aaron Courville. 2016. Deep Learning. The MIT Press.

Syllabus of lectures

  1. The tasks of classification and pattern recognition, basic schema of a classifier, data sets and evaluation
  2. Probabilistic distributions, statistical pattern recognition
  3. Generative and discriminative models
  4. Multivariate Gaussian distribution, Maximum Likelihood estimation,
  5. Gaussian Mixture Model (GMM), Expectation Maximization (EM)
  6. Feature extraction, Mel-frequency cepstral coefficients.
  7. Application of the statistical models in speech and image processing.
  8. Linear classifiers, perceptron
  9. Gaussian Linear Classifier, Logistic regression
  10. Support Vector Machines (SVM), kernel functions
  11. Neural networks - feed-forward, convolutional and recurrent
  12. Hidden Markov Models (HMM) and their application to speech recognition
  13. Project presentation

Syllabus - others, projects and individual work of students

  • Individually assigned projects

Progress assessment

  • Mid-term test - up to 15 points
  • Project - up to 25 points
  • Written final exam - up to 60 points

Controlled instruction

The evaluation includes mid-term test, individual project, and the final exam. The mid-term test does not have a correction option, the final exam has two possible correction terms

Course inclusion in study plans

  • Programme BIT, 2nd year of study, Elective
  • Programme IT-BC-3, field BIT, any year of study, Elective
  • Programme IT-BC-3, field BIT, 2nd year of study, Elective
Back to top