工作流程
计算机科学
可靠性
透明度(行为)
数据科学
相关性(法律)
检查表
人工智能
完备性(序理论)
管理科学
心理学
认识论
工程类
数学
政治学
哲学
计算机安全
数据库
法学
认知心理学
数学分析
作者
Andreas Bender,Nadine Schneider,Marwin Segler,W. Patrick Walters,Ola Engkvist,Tiago Rodrigues
标识
DOI:10.1038/s41570-022-00391-9
摘要
Machine learning (ML) promises to tackle the grand challenges in chemistry and speed up the generation, improvement and/or ordering of research hypotheses. Despite the overarching applicability of ML workflows, one usually finds diverse evaluation study designs. The current heterogeneity in evaluation techniques and metrics leads to difficulty in (or the impossibility of) comparing and assessing the relevance of new algorithms. Ultimately, this may delay the digitalization of chemistry at scale and confuse method developers, experimentalists, reviewers and journal editors. In this Perspective, we critically discuss a set of method development and evaluation guidelines for different types of ML-based publications, emphasizing supervised learning. We provide a diverse collection of examples from various authors and disciplines in chemistry. While taking into account varying accessibility across research groups, our recommendations focus on reporting completeness and standardizing comparisons between tools. We aim to further contribute to improved ML transparency and credibility by suggesting a checklist of retro-/prospective tests and dissecting their importance. We envisage that the wide adoption and continuous update of best practices will encourage an informed use of ML on real-world problems related to the chemical sciences.
科研通智能强力驱动
Strongly Powered by AbleSci AI