计算机科学
变压器
人工智能
编码器
多任务学习
机器学习
Boosting(机器学习)
卷积神经网络
强化学习
计算
任务(项目管理)
算法
操作系统
物理
量子力学
经济
电压
管理
作者
Xiaogang Xu,Hengshuang Zhao,Vibhav Vineet,Ser-Nam Lim,Antonio Torralba
标识
DOI:10.1007/978-3-031-19812-0_18
摘要
In this paper, we explore the advantages of utilizing transformer structures for addressing multi-task learning (MTL). Specifically, we demonstrate that models with transformer structures are more appropriate for MTL than convolutional neural networks (CNNs), and we propose a novel transformer-based architecture named MTFormer for MTL. In the framework, multiple tasks share the same transformer encoder and transformer decoder, and lightweight branches are introduced to harvest task-specific outputs, which increases the MTL performance and reduces the time-space complexity. Furthermore, information from different task domains can benefit each other, and we conduct cross-task reasoning. We propose a cross-task attention mechanism for further boosting the MTL results. The cross-task attention mechanism brings little parameters and computations while introducing extra performance improvements. Besides, we design a self-supervised cross-task contrastive learning algorithm for further boosting the MTL performance. Extensive experiments are conducted on two multi-task learning datasets, on which MTFormer achieves state-of-the-art results with limited network parameters and computations. It also demonstrates significant superiorities for few-shot learning and zero-shot learning.
科研通智能强力驱动
Strongly Powered by AbleSci AI