Course syllabus for Artificial neural networks

Course syllabus adopted 2021-02-26 by Head of Programme (or corresponding).

Overview

  • Swedish nameArtificiella neurala nätverk
  • CodeFFR135
  • Credits7.5 Credits
  • OwnerMPCAS
  • Education cycleSecond-cycle
  • Main field of studyBioengineering, Chemical Engineering, Engineering Physics
  • DepartmentPHYSICS
  • GradingTH - Pass with distinction (5), Pass with credit (4), Pass (3), Fail

Course round 1

  • Teaching language English
  • Application code 11121
  • Maximum participants200
  • Block schedule
  • Open for exchange studentsYes

Credit distribution

0100 Examination 7.5 c
Grading: TH
7.5 c
  • 25 Okt 2021 am L
  • 04 Jan 2022 pm J
  • 18 Aug 2022 pm J

In programmes

Examiner

Go to coursepage (Opens in new tab)

Eligibility

General entry requirements for Master's level (second cycle)
Applicants enrolled in a programme at Chalmers where the course is included in the study programme are exempted from fulfilling the requirements above.

Specific entry requirements

English 6 (or by other approved means with the equivalent proficiency level)
Applicants enrolled in a programme at Chalmers where the course is included in the study programme are exempted from fulfilling the requirements above.

Course specific prerequisites

Analysis in one variable, linear algebra, basic skills in analysis in several variables, and programming.

Aim

This course describes how neural networks are used in machine learning. Neural networks are distributed computational models inspired by the structure of the human brain, consisting of many simple processing elements that are connected in a network. Neural networks have revolutionised how we solve important problems in the engineering sciences, such as image analysis (object recognition and location), prediction, and control. The course gives an overview and a basic understanding of currently used neural-network algorithms, and exhibits similarities as well as differences between these methods. The main emphasis of this introductory course is on three connected topics: recurrent (Hopfield) networks, supervised learning with deep neural networks, and unsupervised learning (reinforcement learning). The goal is to explain how and why the algorithms work, when and how they fail, how to program the standard methods from scratch, and how to use packages that allow to easily set up and to efficiently run larger networks.

Learning outcomes (after completion of the course the student should be able to)

-distinguish between supervised and unsupervised learning, explain the key principles of the corresponding algorithms, understand differences and similarities
-understand under which circumstances neural-net algorithms are the method of choice
-understand and explain strengths and weaknesses of the neural-net algorithms
-implement the algorithms introduced in class on a computer, both from scratch and using neural-net packages
-interpret the results of computer simulations and communicate conclusions in a clear, logical, and concise fashion
-understand the historical development of the field of machine learning with neural networks
-reflect ethical questions posed by machine learning, as well as possible risks

Content

The course is based on Machine learning with neural networks https://arxiv.org/abs/1901.05639. 

1. Statistical mechanics of neural nets
McCulloch-Pitts neurons, Hopfield nets, stochastic optimisation, Boltzmann machines
2. Deep learning
Perceptrons, backpropagation, stochastic gradient descent, deep learning, recurrent nets
3. Unsupervised learning
Hebbian learning, radial basis-function nets, reinforcement learning

Organisation

Lectures

Homework problems
Programming with programming language of choice (commonly matlab or python). We use OpenTA for the homework problems.

Exercise classes
Homework problems and exam questions

Guest lectures
From research and/or industry, possibilities for MSc theses

Short instruction videos

Literature

Course book


B. Mehlig Machine learning with neural networks https://arxiv.org/abs/1901.05639


Additional references


I. Goodfellow, Y. Bengio & A. Courville, Deep Learning https://www.deeplearningbook.org

J. Hertz, A. Krogh & R. G. Palmer, Introduction to the theory of neural computation, Addison-Wesely, Redwood City (1991).


S. Haykin, Neural Networks: a comprehensive foundation, 2nd ed., Prentice Hall, New Jersey (1999)


R.S. Sutton & A. Barto, Reinforcement learning: An Introduction, 2nd ed., MIT Press, http://www.incompleteideas.net/book/the-book-2nd.html

Examination including compulsory elements

The final grade is based on homework assignments (50%) as well as on a written examination (50%). To pass the course one must obtain more than a given minimum number of points in the written exam.

The course examiner may assess individual students in other ways than what is stated above if there are special reasons for doing so, for example if a student has a decision from Chalmers on educational support due to disability.