Neural scaling of deep chemical models

Massive scale, in terms of both data availability and computation, enables important breakthroughs in key application areas of deep learning such as natural language processing and computer vision. There is emerging evidence that scale may be a key ingredient in scientific deep learning, but the imp...

Full description

Bibliographic Details
Main Authors: Frey, Nathan C, Soklaski, Ryan, Axelrod, Simon, Samsi, Siddharth, Gómez-Bombarelli, Rafael, Coley, Connor W, Gadepally, Vijay
Other Authors: Lincoln Laboratory
Format: Article
Language:English
Published: Springer Science and Business Media LLC 2025
Online Access:https://hdl.handle.net/1721.1/158195
_version_ 1824458152219770880
author Frey, Nathan C
Soklaski, Ryan
Axelrod, Simon
Samsi, Siddharth
Gómez-Bombarelli, Rafael
Coley, Connor W
Gadepally, Vijay
author2 Lincoln Laboratory
author_facet Lincoln Laboratory
Frey, Nathan C
Soklaski, Ryan
Axelrod, Simon
Samsi, Siddharth
Gómez-Bombarelli, Rafael
Coley, Connor W
Gadepally, Vijay
author_sort Frey, Nathan C
collection MIT
description Massive scale, in terms of both data availability and computation, enables important breakthroughs in key application areas of deep learning such as natural language processing and computer vision. There is emerging evidence that scale may be a key ingredient in scientific deep learning, but the importance of physical priors in scientific domains makes the strategies and benefits of scaling uncertain. Here we investigate neural-scaling behaviour in large chemical models by varying model and dataset sizes over many orders of magnitude, studying models with over one billion parameters, pre-trained on datasets of up to ten million datapoints. We consider large language models for generative chemistry and graph neural networks for machine-learned interatomic potentials. We investigate the interplay between physical priors and scale and discover empirical neural-scaling relations for language models in chemistry with a scaling exponent of 0.17 for the largest dataset size considered, and a scaling exponent of 0.26 for equivariant graph neural network interatomic potentials.
first_indexed 2025-02-19T04:21:21Z
format Article
id mit-1721.1/158195
institution Massachusetts Institute of Technology
language English
last_indexed 2025-02-19T04:21:21Z
publishDate 2025
publisher Springer Science and Business Media LLC
record_format dspace
spelling mit-1721.1/1581952025-02-11T21:05:34Z Neural scaling of deep chemical models Frey, Nathan C Soklaski, Ryan Axelrod, Simon Samsi, Siddharth Gómez-Bombarelli, Rafael Coley, Connor W Gadepally, Vijay Lincoln Laboratory Massachusetts Institute of Technology. Department of Materials Science and Engineering Massachusetts Institute of Technology. Department of Chemical Engineering Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science Massive scale, in terms of both data availability and computation, enables important breakthroughs in key application areas of deep learning such as natural language processing and computer vision. There is emerging evidence that scale may be a key ingredient in scientific deep learning, but the importance of physical priors in scientific domains makes the strategies and benefits of scaling uncertain. Here we investigate neural-scaling behaviour in large chemical models by varying model and dataset sizes over many orders of magnitude, studying models with over one billion parameters, pre-trained on datasets of up to ten million datapoints. We consider large language models for generative chemistry and graph neural networks for machine-learned interatomic potentials. We investigate the interplay between physical priors and scale and discover empirical neural-scaling relations for language models in chemistry with a scaling exponent of 0.17 for the largest dataset size considered, and a scaling exponent of 0.26 for equivariant graph neural network interatomic potentials. 2025-02-11T21:05:33Z 2025-02-11T21:05:33Z 2023 2025-02-11T20:58:27Z Article http://purl.org/eprint/type/JournalArticle https://hdl.handle.net/1721.1/158195 Frey, N.C., Soklaski, R., Axelrod, S. et al. Neural scaling of deep chemical models. Nat Mach Intell 5, 1297–1305 (2023). en 10.1038/s42256-023-00740-3 Nature Machine Intelligence Creative Commons Attribution https://creativecommons.org/licenses/by/4.0/ application/pdf Springer Science and Business Media LLC Springer Science and Business Media LLC
spellingShingle Frey, Nathan C
Soklaski, Ryan
Axelrod, Simon
Samsi, Siddharth
Gómez-Bombarelli, Rafael
Coley, Connor W
Gadepally, Vijay
Neural scaling of deep chemical models
title Neural scaling of deep chemical models
title_full Neural scaling of deep chemical models
title_fullStr Neural scaling of deep chemical models
title_full_unstemmed Neural scaling of deep chemical models
title_short Neural scaling of deep chemical models
title_sort neural scaling of deep chemical models
url https://hdl.handle.net/1721.1/158195
work_keys_str_mv AT freynathanc neuralscalingofdeepchemicalmodels
AT soklaskiryan neuralscalingofdeepchemicalmodels
AT axelrodsimon neuralscalingofdeepchemicalmodels
AT samsisiddharth neuralscalingofdeepchemicalmodels
AT gomezbombarellirafael neuralscalingofdeepchemicalmodels
AT coleyconnorw neuralscalingofdeepchemicalmodels
AT gadepallyvijay neuralscalingofdeepchemicalmodels