This lecture offers a journey through recent approaches towards a deeper theo-retical understanding of deep neural networks, focussing on questions such as properties of deep nets near initialization, mean field theory, complexity results for deep networks, properties of loss landscapes and gradient learning, infinite width limit of learning, neural tangent kernel, and renormalization group ap-proaches for the non-gaussian regime. Many results will be seen to stem from an analysis of certain limiting cases of neural network architectures (linear, vanishing depth/width ratio) that provide insights about properties of real net-works that are "close" to such cases.
The lecture will proceed along a "theory backbone" provided in the recent book "The principles of deep learning theory" by Roberts, Yaida and Hanin (2021), exploiting methods from theoretical physics for analyzing properties of deep networks which reveal striking analogies with phenomena in physical many-particle systems. This backbone will be complemented with selected papers that deepen the topics listed above further.
Prerequisites are a solid background in linear algebra, multivariate calculus and rudimentary probability theory.
Frequency | Weekday | Time | Format / Place | Period |
---|
Module | Course | Requirements | |
---|---|---|---|
39-M-Inf-AI-adv-foc Advanced Artificial Intelligence (focus) | Advanced Artificial Intelligence (focus): Vorlesung | Graded examination
|
Student information |
39-M-Inf-VML Vertiefung Maschinelles Lernen | Vertiefung Maschinelles Lernen | Student information |
The binding module descriptions contain further information, including specifications on the "types of assignments" students need to complete. In cases where a module description mentions more than one kind of assignment, the respective member of the teaching staff will decide which task(s) they assign the students.
A corresponding course offer for this course already exists in the e-learning system. Teaching staff can store materials relating to teaching courses there: