标杆管理
苯胺
计算机科学
反应条件
强化学习
多样性(控制论)
组合化学
生化工程
化学
人工智能
催化作用
有机化学
工程类
营销
业务
作者
Jason Y. Wang,Jason M. Stevens,Stavros K. Kariofillis,Mai-Jan Tom,Dung L. Golden,Jun Li,José E. Tábora,Marvin Parasram,Benjamin J. Shields,David N. Primer,Bo Hao,David Del Valle,Stacey DiSomma,A.H. Furman,Greg Zipp,Sergey Melnikov,James Paulson,Abigail G. Doyle
出处
期刊:Nature
[Springer Nature]
日期:2024-02-28
卷期号:626 (8001): 1025-1033
被引量:12
标识
DOI:10.1038/s41586-024-07021-y
摘要
Reaction conditions that are generally applicable to a wide variety of substrates are highly desired, especially in the pharmaceutical and chemical industries1–6. Although many approaches are available to evaluate the general applicability of developed conditions, a universal approach to efficiently discover these conditions during optimizations is rare. Here we report the design, implementation and application of reinforcement learning bandit optimization models7–10 to identify generally applicable conditions by efficient condition sampling and evaluation of experimental feedback. Performance benchmarking on existing datasets statistically showed high accuracies for identifying general conditions, with up to 31% improvement over baselines that mimic state-of-the-art optimization approaches. A palladium-catalysed imidazole C–H arylation reaction, an aniline amide coupling reaction and a phenol alkylation reaction were investigated experimentally to evaluate use cases and functionalities of the bandit optimization model in practice. In all three cases, the reaction conditions that were most generally applicable yet not well studied for the respective reaction were identified after surveying less than 15% of the expert-designed reaction space. Bandit optimization models are used to identify generally applicable conditions by efficient condition sampling and evaluation of experimental feedback.
科研通智能强力驱动
Strongly Powered by AbleSci AI