Transformer graph variational autoencoder for generative molecular design

In the field of drug discovery, the generation of new molecules with desirable properties remains a critical challenge. Traditional methods often rely on simplified molecular input line entry system representations for molecular input data, which can limit the diversity and novelty of generated mole...

Full description

Saved in:
Bibliographic Details
Published in:Biophysical journal Vol. 124; no. 22; p. 3867
Main Authors: Nguyen, Trieu, Karolak, Aleksandra
Format: Journal Article
Language:English
Published: United States 18.11.2025
Subjects:
ISSN:1542-0086, 1542-0086
Online Access:Get more information
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:In the field of drug discovery, the generation of new molecules with desirable properties remains a critical challenge. Traditional methods often rely on simplified molecular input line entry system representations for molecular input data, which can limit the diversity and novelty of generated molecules. To address this, we present the transformer graph variational autoencoder (TGVAE), an innovative AI model that employs molecular graphs as input data, thus capturing the complex structural relationships within molecules more effectively than string models. To enhance molecular generation capabilities, TGVAE combines a transformer, graph neural network (GNN), and VAE. Additionally, we address common issues like over-smoothing in training GNNs and posterior collapse in VAEs to ensure robust training and improve the generation of chemically valid and diverse molecular structures. Our results demonstrate that TGVAE outperforms existing approaches, generating a larger collection of diverse molecules and discovering structures that were previously unexplored. This advancement not only brings more possibilities for drug discovery but also sets a new level for the use of AI in molecular generation.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
ISSN:1542-0086
1542-0086
DOI:10.1016/j.bpj.2025.01.022