J. Berner, P. Grohs, G. Kutyniok, и P. Petersen. (2021)cite arxiv:2105.04026Comment: This review paper will appear as a book chapter in the book "Theory of Deep Learning" by Cambridge University Press.
Аннотация
We describe the new field of mathematical analysis of deep learning. This
field emerged around a list of research questions that were not answered within
the classical framework of learning theory. These questions concern: the
outstanding generalization power of overparametrized neural networks, the role
of depth in deep architectures, the apparent absence of the curse of
dimensionality, the surprisingly successful optimization performance despite
the non-convexity of the problem, understanding what features are learned, why
deep architectures perform exceptionally well in physical problems, and which
fine aspects of an architecture affect the behavior of a learning task in which
way. We present an overview of modern approaches that yield partial answers to
these questions. For selected approaches, we describe the main ideas in more
detail.
%0 Generic
%1 berner2021modern
%A Berner, Julius
%A Grohs, Philipp
%A Kutyniok, Gitta
%A Petersen, Philipp
%D 2021
%K math
%T The Modern Mathematics of Deep Learning
%U http://arxiv.org/abs/2105.04026
%X We describe the new field of mathematical analysis of deep learning. This
field emerged around a list of research questions that were not answered within
the classical framework of learning theory. These questions concern: the
outstanding generalization power of overparametrized neural networks, the role
of depth in deep architectures, the apparent absence of the curse of
dimensionality, the surprisingly successful optimization performance despite
the non-convexity of the problem, understanding what features are learned, why
deep architectures perform exceptionally well in physical problems, and which
fine aspects of an architecture affect the behavior of a learning task in which
way. We present an overview of modern approaches that yield partial answers to
these questions. For selected approaches, we describe the main ideas in more
detail.
@misc{berner2021modern,
abstract = {We describe the new field of mathematical analysis of deep learning. This
field emerged around a list of research questions that were not answered within
the classical framework of learning theory. These questions concern: the
outstanding generalization power of overparametrized neural networks, the role
of depth in deep architectures, the apparent absence of the curse of
dimensionality, the surprisingly successful optimization performance despite
the non-convexity of the problem, understanding what features are learned, why
deep architectures perform exceptionally well in physical problems, and which
fine aspects of an architecture affect the behavior of a learning task in which
way. We present an overview of modern approaches that yield partial answers to
these questions. For selected approaches, we describe the main ideas in more
detail.},
added-at = {2021-06-11T20:00:55.000+0200},
author = {Berner, Julius and Grohs, Philipp and Kutyniok, Gitta and Petersen, Philipp},
biburl = {https://www.bibsonomy.org/bibtex/29dbf16671b953bcd6bfa48ab7208c1a5/new4761},
description = {The Modern Mathematics of Deep Learning},
interhash = {d546c743b2f7274343cd42e542e257a1},
intrahash = {9dbf16671b953bcd6bfa48ab7208c1a5},
keywords = {math},
note = {cite arxiv:2105.04026Comment: This review paper will appear as a book chapter in the book "Theory of Deep Learning" by Cambridge University Press},
timestamp = {2021-06-11T20:00:55.000+0200},
title = {The Modern Mathematics of Deep Learning},
url = {http://arxiv.org/abs/2105.04026},
year = 2021
}