The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale's 18th problem
View / Open Files
Authors
Colbrook, Matthew J
Antun, Vegard
Hansen, Anders C
Publication Date
2022Journal Title
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA
ISSN
0027-8424
Publisher
National Academy of Sciences
Volume
119
Issue
12
Language
eng
Type
Article
This Version
VoR
Metadata
Show full item recordCitation
Colbrook, M. J., Antun, V., & Hansen, A. C. (2022). The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale's 18th problem. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 119 (12) https://doi.org/10.1073/pnas.2107151119
Abstract
Deep learning (DL) has had unprecedented success and is now entering
scientific computing with full force. However, current DL methods typically
suffer from instability, even when universal approximation properties guarantee
the existence of stable neural networks (NNs). We address this paradox by
demonstrating basic well-conditioned problems in scientific computing where one
can prove the existence of NNs with great approximation qualities, however,
there does not exist any algorithm, even randomised, that can train (or
compute) such a NN. For any positive integers $K > 2$ and $L$, there are cases
where simultaneously: (a) no randomised training algorithm can compute a NN
correct to $K$ digits with probability greater than $1/2$, (b) there exists a
deterministic training algorithm that computes a NN with $K-1$ correct digits,
but any such (even randomised) algorithm needs arbitrarily many training data,
(c) there exists a deterministic training algorithm that computes a NN with
$K-2$ correct digits using no more than $L$ training samples. These results
imply a classification theory describing conditions under which (stable) NNs
with a given accuracy can be computed by an algorithm. We begin this theory by
establishing sufficient conditions for the existence of algorithms that compute
stable NNs in inverse problems. We introduce Fast Iterative REstarted NETworks
(FIRENETs), which we both prove and numerically verify are stable. Moreover, we
prove that only $\mathcal{O}(|\log(\epsilon)|)$ layers are needed for an
$\epsilon$-accurate solution to the inverse problem.
Keywords
stability and accuracy, AI and deep learning, inverse problems, Smale's 18th problem, solvability complexity index hierarchy
Sponsorship
Leverhulme Prize (n/a)
Royal Society (n/a)
Trinity College, University of Cambridge (n/a)
Identifiers
35294283, PMC8944871
External DOI: https://doi.org/10.1073/pnas.2107151119
This record's URL: https://www.repository.cam.ac.uk/handle/1810/336157
Rights
Attribution-NonCommercial-NoDerivatives 4.0 International
Licence URL: https://creativecommons.org/licenses/by-nc-nd/4.0/
Statistics
Total file downloads (since January 2020). For more information on metrics see the
IRUS guide.
Recommended or similar items
The current recommendation prototype on the Apollo Repository will be turned off on 03 February 2023. Although the pilot has been fruitful for both parties, the service provider IKVA is focusing on horizon scanning products and so the recommender service can no longer be supported. We recognise the importance of recommender services in supporting research discovery and are evaluating offerings from other service providers. If you would like to offer feedback on this decision please contact us on: support@repository.cam.ac.uk