计算机科学
推荐系统
变压器
点积
机器学习
人工智能
等价(形式语言)
水准点(测量)
计算复杂性理论
理论计算机科学
算法
数学
物理
离散数学
电压
量子力学
地理
大地测量学
几何学
作者
Langming Liu,Liu Cai,Chi Zhang,Xiangyu Zhao,Jingtong Gao,Wanyu Wang,Yifu Lv,Wenqi Fan,Yiqi Wang,Ming He,Zitao Liu,Qing Li
标识
DOI:10.1145/3539618.3591717
摘要
Transformer models have achieved remarkable success in sequential recommender systems (SRSs). However, computing the attention matrix in traditional dot-product attention mechanisms results in a quadratic complexity with sequence lengths, leading to high computational costs for long-term sequential recommendation. Motivated by the above observation, we propose a novel L2-Normalized Linear Attention for the Transformer-based Sequential Recommender Systems (LinRec), which theoretically improves efficiency while preserving the learning capabilities of the traditional dot-product attention. Specifically, by thoroughly examining the equivalence conditions of efficient attention mechanisms, we show that LinRec possesses linear complexity while preserving the property of attention mechanisms. In addition, we reveal its latent efficiency properties by interpreting the proposed LinRec mechanism through a statistical lens. Extensive experiments are conducted based on two public benchmark datasets, demonstrating that the combination of LinRec and Transformer models achieves comparable or even superior performance than state-of-the-art Transformer-based SRS models while significantly improving time and memory efficiency. The implementation code is available online at https://github.com/Applied-Machine-Learning-Lab/LinRec.>
科研通智能强力驱动
Strongly Powered by AbleSci AI