Expressivity of Neural Networks

Type

Master thesis / Bachelor thesis

Prerequisites

  • Knowledge of Python and deep learning frameworks (TensorFlow or PyTorch)
  • Preferably background in functional analysis

Description

Deep learning models have achieved tremendous success in many application areas. This also raises interesting theoretical questions such as: Why can deep learning models adapt to widely different tasks? Are we able to describe their expressive power? A longstanding result in this direction is the Universal Approximation Theorem, which states that neural networks with only two layers can, under weak conditions, approximate any continuous function on a compact set up to arbitrary precision. However, this is only a starting point for the above questions. A common goal is to relate the approximation accuracy of neural networks to their complexity. It turns out that for many well-known classical function spaces, one can derive upper and lower complexity bounds in terms of the number of weights, neurons, and layers. Further exciting lines of research try to explain the benefits of deep neural networks over shallow ones and study how neural architectural design choices influence expressivity. These examples are among the many interesting and open research questions within the field of neural network expressivity.

References