Backprop as Functor: A compositional perspective on supervised learning

A supervised learning algorithm searches over a set of functions A→B parametrised by a space P to find the best approximation to some ideal function f:A→B. It does this by taking examples (a,f(a))∈A×B, and updating the parameter according to some rule. We define a category where these update rules m...

Full description

Bibliographic Details
Main Authors: Fong, Brendan C, Spivak, David I, Tuyeras, Remy V
Other Authors: Massachusetts Institute of Technology. Department of Mathematics
Format: Article
Language:en_US
Published: Association for Computing Machinery 2020
Online Access:https://hdl.handle.net/1721.1/123513
Description
Summary:A supervised learning algorithm searches over a set of functions A→B parametrised by a space P to find the best approximation to some ideal function f:A→B. It does this by taking examples (a,f(a))∈A×B, and updating the parameter according to some rule. We define a category where these update rules may be composed, and show that gradient descent---with respect to a fixed step size and an error function satisfying a certain property---defines a monoidal functor from a category of parametrised functions to this category of update rules. This provides a structural perspective on backpropagation, as well as a broad generalisation of neural networks.