计算机科学
任务(项目管理)
遗忘
航程(航空)
人工智能
一套
多任务学习
分布式计算
语言学
哲学
材料科学
管理
考古
经济
复合材料
历史
作者
Matthew Wallingford,Hao Li,Alessandro Achille,Avinash Ravichandran,Charless C. Fowlkes,Rahul Bhotika,Stefano Soatto
标识
DOI:10.1109/cvpr52688.2022.00741
摘要
Adapting pre-trained models with broad capabilities has become standard practice for learning a wide range of downstream tasks. The typical approach of fine-tuning different models for each task is performant, but incurs a substantial memory cost. To efficiently learn multiple down-stream tasks we introduce Task Adaptive Parameter Sharing (TAPS), a simple method for tuning a base model to a new task by adaptively modifying a small, task-specific subset of layers. This enables multi-task learning while minimizing the resources used and avoids catastrophic forgetting and competition between tasks. TAPS solves a joint optimization problem which determines both the layers that are shared with the base model and the value of the task-specific weights. Further, a sparsity penalty on the number of active layers promotes weight sharing with the base model. Compared to other methods, TAPS retains a high accuracy on the target tasks while still introducing only a small number of task-specific parameters. Moreover, TAPS is agnostic to the particular architecture used and requires only minor changes to the training scheme. We evaluate our method on a suite of fine-tuning tasks and architectures (ResNet, DenseNet, ViT) and show that it achieves state-of-the-art performance while being simple to implement.
科研通智能强力驱动
Strongly Powered by AbleSci AI