计算机科学
多任务学习
任务(项目管理)
人工智能
机器学习
学习迁移
集合(抽象数据类型)
领域(数学)
特征(语言学)
一般化
程序设计语言
数学
语言学
数学分析
哲学
经济
管理
纯数学
作者
Ke Ding,Xin Dong,Yong He,Lei Cheng,Chilin Fu,Zhaoxin Huan,Hai Li,Tan Yan,Liang Zhang,Xiaolu Zhang,Linjian Mo
标识
DOI:10.1145/3404835.3463022
摘要
Multi-task learning(MTL) is an open and challenging problem in various real-world applications. The typical way of conducting multi-task learning is establishing some global parameter sharing mechanism across all tasks or assigning each task an individual set of parameters with cross-connections between tasks. However, for most existing approaches, all tasks just thoroughly or proportionally share all the features without distinguishing the helpfulness of them. By that, some tasks would be intervened by the unhelpful features that are useful for other tasks, leading to undesired negative transfer between tasks. In this paper, we design a novel architecture named the Multiple-level Sparse Sharing Model (MSSM), which can learn features selectively and share knowledge across all tasks efficiently. MSSM first employs a field-level sparse connection module (FSCM) to enable much more expressive combinations of feature fields to be learned for generalization across tasks while still allowing for task-specific features to be customized for each task. Furthermore, a cell-level sparse sharing module (CSSM) can recognize the sharing pattern through a set of coding variables that selectively choose which cells to route for a given task. Extensive experimental results on several real-world datasets show that MSSM outperforms SOTA models significantly in terms of AUC and LogLoss metrics.
科研通智能强力驱动
Strongly Powered by AbleSci AI