可解释性
山崩
口译(哲学)
地图学
仰角(弹道)
领域(数学)
地理
地质学
机器学习
遥感
人工智能
计算机科学
数学
地貌学
几何学
纯数学
程序设计语言
作者
Xinzhi Zhou,Haijia Wen,Ziwei Li,Hui Zhang,Wengang Zhang
标识
DOI:10.1080/10106049.2022.2076928
摘要
The machine-learning “black box” models, which lack interpretability, have limited application in landslide susceptibility mapping. To interpret the black-box models, some interpretable machine learning algorithms have been proposed recently. Among them is SHaply Additive ExPlanation (SHAP), which has attracted much attention because of its ease of operation and comprehensiveness. In this study, a novel interpretable model based on SHAP and XGBoost is proposed to interpret landslides susceptibility evaluation at global and local levels. The established evaluation model provided 0.75 accuracy and 0.83 AUC value for the test sets. The global interpretation shows that the peak rainfall intensity and elevation are the dominant factors that influence the occurrence of landslides in the study area. The combination of local interpretation and field investigations can provide a comprehensive framework for evaluating designated landslides, and it can also be used as a reference for preventing and managing the hazards of landslides.
科研通智能强力驱动
Strongly Powered by AbleSci AI