化学
化学空间
计算机科学
变压器
训练集
人工智能
药物发现
赫尔格
小分子
机器学习
计算生物学
化学
生物
工程类
生物化学
钾通道
生物物理学
电压
电气工程
作者
Emma Tysinger,K. Brajesh,Anton V. Sinitskiy
标识
DOI:10.1021/acs.jcim.2c01618
摘要
Meaningful exploration of the chemical space of druglike molecules in drug design is a highly challenging task due to a combinatorial explosion of possible modifications of molecules. In this work, we address this problem with transformer models, a type of machine learning (ML) model originally developed for machine translation. By training transformer models on pairs of similar bioactive molecules from the public ChEMBL data set, we enable them to learn medicinal-chemistry-meaningful, context-dependent transformations of molecules, including those absent from the training set. By retrospective analysis on the performance of transformer models on ChEMBL subsets of ligands binding to COX2, DRD2, or HERG protein targets, we demonstrate that the models can generate structures identical or highly similar to most active ligands, despite the models having not seen any ligands active against the corresponding protein target during training. Our work demonstrates that human experts working on hit expansion in drug design can easily and quickly employ transformer models, originally developed to translate texts from one natural language to another, to "translate" from known molecules active against a given protein target to novel molecules active against the same target.
科研通智能强力驱动
Strongly Powered by AbleSci AI