Generative Chemical Transformer: Neural Machine Learning of Molecular Geometric Structures from Chemical Language via Attention

Hyunseung Kim, Jonggeol Na, Won Bo Lee

Research output: Contribution to journalReview articlepeer-review

28 Scopus citations

Abstract

Discovering new materials better suited to specific purposes is an important issue in improving the quality of human life. Here, a neural network that creates molecules that meet some desired multiple target conditions based on a deep understanding of chemical language is proposed (generative chemical Transformer, GCT). The attention mechanism in GCT allows a deeper understanding of molecular structures beyond the limitations of chemical language itself which cause semantic discontinuity by paying attention to characters sparsely. The significance of language models for inverse molecular design problems is investigated by quantitatively evaluating the quality of the generated molecules. GCT generates highly realistic chemical strings that satisfy both chemical and linguistic grammar rules. Molecules parsed from the generated strings simultaneously satisfy the multiple target properties and vary for a single condition set. These advances will contribute to improving the quality of human life by accelerating the process of desired material discovery.

Original languageEnglish
Pages (from-to)5804-5814
Number of pages11
JournalJournal of Chemical Information and Modeling
Volume61
Issue number12
DOIs
StatePublished - 27 Dec 2021

Bibliographical note

Publisher Copyright:
© 2021 American Chemical Society.

Fingerprint

Dive into the research topics of 'Generative Chemical Transformer: Neural Machine Learning of Molecular Geometric Structures from Chemical Language via Attention'. Together they form a unique fingerprint.

Cite this