Molecule generation using transformers and policy gradient reinforcement learning

Abstract Generating novel valid molecules is often a difficult task, because the vast chemical space relies on the intuition of experienced chemists. In recent years, deep learning models have helped accelerate this process. These advanced models can also help identify suitable molecules for disease...

Full description

Bibliographic Details
Main Authors: Eyal Mazuz, Guy Shtar, Bracha Shapira, Lior Rokach
Format: Article
Language:English
Published: Nature Portfolio 2023-05-01
Series:Scientific Reports
Online Access:https://doi.org/10.1038/s41598-023-35648-w
_version_ 1797811478403743744
author Eyal Mazuz
Guy Shtar
Bracha Shapira
Lior Rokach
author_facet Eyal Mazuz
Guy Shtar
Bracha Shapira
Lior Rokach
author_sort Eyal Mazuz
collection DOAJ
description Abstract Generating novel valid molecules is often a difficult task, because the vast chemical space relies on the intuition of experienced chemists. In recent years, deep learning models have helped accelerate this process. These advanced models can also help identify suitable molecules for disease treatment. In this paper, we propose Taiga, a transformer-based architecture for the generation of molecules with desired properties. Using a two-stage approach, we first treat the problem as a language modeling task of predicting the next token, using SMILES strings. Then, we use reinforcement learning to optimize molecular properties such as QED. This approach allows our model to learn the underlying rules of chemistry and more easily optimize for molecules with desired properties. Our evaluation of Taiga, which was performed with multiple datasets and tasks, shows that Taiga is comparable to, or even outperforms, state-of-the-art baselines for molecule optimization, with improvements in the QED ranging from 2 to over 20 percent. The improvement was demonstrated both on datasets containing lead molecules and random molecules. We also show that with its two stages, Taiga is capable of generating molecules with higher biological property scores than the same model without reinforcement learning.
first_indexed 2024-03-13T07:24:10Z
format Article
id doaj.art-43421964cb2b49f9889efbc7f325dfb3
institution Directory Open Access Journal
issn 2045-2322
language English
last_indexed 2024-03-13T07:24:10Z
publishDate 2023-05-01
publisher Nature Portfolio
record_format Article
series Scientific Reports
spelling doaj.art-43421964cb2b49f9889efbc7f325dfb32023-06-04T11:28:45ZengNature PortfolioScientific Reports2045-23222023-05-0113111110.1038/s41598-023-35648-wMolecule generation using transformers and policy gradient reinforcement learningEyal Mazuz0Guy Shtar1Bracha Shapira2Lior Rokach3Ben-Gurion University of the NegevBen-Gurion University of the NegevBen-Gurion University of the NegevBen-Gurion University of the NegevAbstract Generating novel valid molecules is often a difficult task, because the vast chemical space relies on the intuition of experienced chemists. In recent years, deep learning models have helped accelerate this process. These advanced models can also help identify suitable molecules for disease treatment. In this paper, we propose Taiga, a transformer-based architecture for the generation of molecules with desired properties. Using a two-stage approach, we first treat the problem as a language modeling task of predicting the next token, using SMILES strings. Then, we use reinforcement learning to optimize molecular properties such as QED. This approach allows our model to learn the underlying rules of chemistry and more easily optimize for molecules with desired properties. Our evaluation of Taiga, which was performed with multiple datasets and tasks, shows that Taiga is comparable to, or even outperforms, state-of-the-art baselines for molecule optimization, with improvements in the QED ranging from 2 to over 20 percent. The improvement was demonstrated both on datasets containing lead molecules and random molecules. We also show that with its two stages, Taiga is capable of generating molecules with higher biological property scores than the same model without reinforcement learning.https://doi.org/10.1038/s41598-023-35648-w
spellingShingle Eyal Mazuz
Guy Shtar
Bracha Shapira
Lior Rokach
Molecule generation using transformers and policy gradient reinforcement learning
Scientific Reports
title Molecule generation using transformers and policy gradient reinforcement learning
title_full Molecule generation using transformers and policy gradient reinforcement learning
title_fullStr Molecule generation using transformers and policy gradient reinforcement learning
title_full_unstemmed Molecule generation using transformers and policy gradient reinforcement learning
title_short Molecule generation using transformers and policy gradient reinforcement learning
title_sort molecule generation using transformers and policy gradient reinforcement learning
url https://doi.org/10.1038/s41598-023-35648-w
work_keys_str_mv AT eyalmazuz moleculegenerationusingtransformersandpolicygradientreinforcementlearning
AT guyshtar moleculegenerationusingtransformersandpolicygradientreinforcementlearning
AT brachashapira moleculegenerationusingtransformersandpolicygradientreinforcementlearning
AT liorrokach moleculegenerationusingtransformersandpolicygradientreinforcementlearning