Rational neural networks

We consider neural networks with rational activation functions. The choice of the nonlinear activation function in deep learning architectures is crucial and heavily impacts the performance of a neural network. We establish optimal bounds in terms of network complexity and prove that rational neural...

Full description

Bibliographic Details
Main Authors: Boullé, N, Nakatsukasa, Y, Townsend, A
Format: Conference item
Language:English
Published: Conference on Neural Information Processing Systems 2020
_version_ 1797067623866826752
author Boullé, N
Nakatsukasa, Y
Townsend, A
author_facet Boullé, N
Nakatsukasa, Y
Townsend, A
author_sort Boullé, N
collection OXFORD
description We consider neural networks with rational activation functions. The choice of the nonlinear activation function in deep learning architectures is crucial and heavily impacts the performance of a neural network. We establish optimal bounds in terms of network complexity and prove that rational neural networks approximate smooth functions more efficiently than ReLU networks with exponentially smaller depth. The flexibility and smoothness of rational activation functions make them an attractive alternative to ReLU, as we demonstrate with numerical experiments.
first_indexed 2024-03-06T21:59:00Z
format Conference item
id oxford-uuid:4df37d2b-1104-4ead-955f-02c95ed9a388
institution University of Oxford
language English
last_indexed 2024-03-06T21:59:00Z
publishDate 2020
publisher Conference on Neural Information Processing Systems
record_format dspace
spelling oxford-uuid:4df37d2b-1104-4ead-955f-02c95ed9a3882022-03-26T15:58:26ZRational neural networksConference itemhttp://purl.org/coar/resource_type/c_c94fuuid:4df37d2b-1104-4ead-955f-02c95ed9a388EnglishSymplectic ElementsConference on Neural Information Processing Systems2020Boullé, NNakatsukasa, YTownsend, AWe consider neural networks with rational activation functions. The choice of the nonlinear activation function in deep learning architectures is crucial and heavily impacts the performance of a neural network. We establish optimal bounds in terms of network complexity and prove that rational neural networks approximate smooth functions more efficiently than ReLU networks with exponentially smaller depth. The flexibility and smoothness of rational activation functions make them an attractive alternative to ReLU, as we demonstrate with numerical experiments.
spellingShingle Boullé, N
Nakatsukasa, Y
Townsend, A
Rational neural networks
title Rational neural networks
title_full Rational neural networks
title_fullStr Rational neural networks
title_full_unstemmed Rational neural networks
title_short Rational neural networks
title_sort rational neural networks
work_keys_str_mv AT boullen rationalneuralnetworks
AT nakatsukasay rationalneuralnetworks
AT townsenda rationalneuralnetworks