On the Complexity of Neural Computation in Superposition
Recent advances in the understanding of neural networks suggest that superposition, the ability of a single neuron to represent multiple features simultaneously, is a key mechanism underlying the computational efficiency of large-scale networks. This paper explores the theoretical foundations of com...
Main Authors: | , |
---|---|
Format: | Article |
Language: | en_US |
Published: |
2024
|
Subjects: | |
Online Access: | https://hdl.handle.net/1721.1/157073 |
_version_ | 1824458165357379584 |
---|---|
author | Adler, Micah Shavit, Nir |
author_facet | Adler, Micah Shavit, Nir |
author_sort | Adler, Micah |
collection | MIT |
description | Recent advances in the understanding of neural networks suggest that superposition, the ability of a single neuron to represent multiple features simultaneously, is a key mechanism underlying the computational efficiency of large-scale networks. This paper explores the theoretical foundations of computing in superposition, focusing on explicit, provably correct algorithms and their efficiency.
We present the first lower bounds showing that for a broad class of problems, including permutations and pairwise logical operations, a neural net- work computing in superposition requires at least Ω(m′ log m′) parameters and Ω(√(m′ log m′)) neurons, where m′ is the number of output features being computed. This implies that any “lottery ticket” sparse sub-network must have at least Ω(m′ log m′ ) parameters no matter what the initial dense network size. Conversely, we show a nearly tight upper bound: logical operations like pair- wise AND can be computed using O(√(m′) log m′) neurons and O(m′ log^2 m′) parameters. There is thus an exponential gap between computing in superposition, the subject of this work, and representing features in superposition, which can require as little as O(log m′) neurons based on the Johnson-Lindenstrauss Lemma.
Our hope is that our results open a path for using complexity theoretic techniques in neural network interpretability research. |
first_indexed | 2025-02-19T04:21:33Z |
format | Article |
id | mit-1721.1/157073 |
institution | Massachusetts Institute of Technology |
language | en_US |
last_indexed | 2025-02-19T04:21:33Z |
publishDate | 2024 |
record_format | dspace |
spelling | mit-1721.1/1570732024-10-01T03:33:22Z On the Complexity of Neural Computation in Superposition Adler, Micah Shavit, Nir superposition neural network neurons complexity Recent advances in the understanding of neural networks suggest that superposition, the ability of a single neuron to represent multiple features simultaneously, is a key mechanism underlying the computational efficiency of large-scale networks. This paper explores the theoretical foundations of computing in superposition, focusing on explicit, provably correct algorithms and their efficiency. We present the first lower bounds showing that for a broad class of problems, including permutations and pairwise logical operations, a neural net- work computing in superposition requires at least Ω(m′ log m′) parameters and Ω(√(m′ log m′)) neurons, where m′ is the number of output features being computed. This implies that any “lottery ticket” sparse sub-network must have at least Ω(m′ log m′ ) parameters no matter what the initial dense network size. Conversely, we show a nearly tight upper bound: logical operations like pair- wise AND can be computed using O(√(m′) log m′) neurons and O(m′ log^2 m′) parameters. There is thus an exponential gap between computing in superposition, the subject of this work, and representing features in superposition, which can require as little as O(log m′) neurons based on the Johnson-Lindenstrauss Lemma. Our hope is that our results open a path for using complexity theoretic techniques in neural network interpretability research. 2024-09-30T15:49:30Z 2024-09-30T15:49:30Z 2024-09-30 Article https://hdl.handle.net/1721.1/157073 en_US Attribution-NonCommercial-NoDerivs 3.0 United States http://creativecommons.org/licenses/by-nc-nd/3.0/us/ application/pdf |
spellingShingle | superposition neural network neurons complexity Adler, Micah Shavit, Nir On the Complexity of Neural Computation in Superposition |
title | On the Complexity of Neural Computation in Superposition |
title_full | On the Complexity of Neural Computation in Superposition |
title_fullStr | On the Complexity of Neural Computation in Superposition |
title_full_unstemmed | On the Complexity of Neural Computation in Superposition |
title_short | On the Complexity of Neural Computation in Superposition |
title_sort | on the complexity of neural computation in superposition |
topic | superposition neural network neurons complexity |
url | https://hdl.handle.net/1721.1/157073 |
work_keys_str_mv | AT adlermicah onthecomplexityofneuralcomputationinsuperposition AT shavitnir onthecomplexityofneuralcomputationinsuperposition |