计算机科学
推荐系统
排名(信息检索)
水准点(测量)
情报检索
局部敏感散列
分类
多样性(控制论)
学习排名
相似性(几何)
编码(集合论)
机器学习
数据挖掘
散列函数
人工智能
哈希表
图像(数学)
集合(抽象数据类型)
程序设计语言
地理
计算机安全
大地测量学
作者
Lei Li,Yongfeng Zhang,Li Chen
标识
DOI:10.1145/3404835.3463248
摘要
Recently, research on explainable recommender systems has drawn much attention from both academia and industry, resulting in a variety of explainable models. As a consequence, their evaluation approaches vary from model to model, which makes it quite difficult to compare the explainability of different models. To achieve a standard way of evaluating recommendation explanations, we provide three benchmark datasets for EXplanaTion RAnking (denoted as EXTRA), on which explainability can be measured by ranking-oriented metrics. Constructing such datasets, however, poses great challenges. First, user-item-explanation triplet interactions are rare in existing recommender systems, so how to find alternatives becomes a challenge. Our solution is to identify nearly identical sentences from user reviews. This idea then leads to the second challenge, i.e., how to efficiently categorize the sentences in a dataset into different groups, since it has quadratic runtime complexity to estimate the similarity between any two sentences. To mitigate this issue, we provide a more efficient method based on Locality Sensitive Hashing (LSH) that can detect near-duplicates in sub-linear time for a given query. Moreover, we make our code publicly available to allow researchers in the community to create their own datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI