可学性
计算机科学
背景(考古学)
数据科学
问责
认知
控制(管理)
人工智能
认知科学
人机交互
心理学
政治学
古生物学
神经科学
法学
生物
作者
Ashraf Abdul,Jo Vermeulen,Danding Wang,Brian Y. Lim,Mohan Kankanhalli
标识
DOI:10.1145/3173574.3174156
摘要
Advances in artificial intelligence, sensors and big data management have far-reaching societal impacts. As these systems augment our everyday lives, it becomes increasing-ly important for people to understand them and remain in control. We investigate how HCI researchers can help to develop accountable systems by performing a literature analysis of 289 core papers on explanations and explaina-ble systems, as well as 12,412 citing papers. Using topic modeling, co-occurrence and network analysis, we mapped the research space from diverse domains, such as algorith-mic accountability, interpretable machine learning, context-awareness, cognitive psychology, and software learnability. We reveal fading and burgeoning trends in explainable systems, and identify domains that are closely connected or mostly isolated. The time is ripe for the HCI community to ensure that the powerful new autonomous systems have intelligible interfaces built-in. From our results, we propose several implications and directions for future research to-wards this goal.
科研通智能强力驱动
Strongly Powered by AbleSci AI