计算机科学
任务(项目管理)
多任务学习
人工智能
适应性
机器学习
任务分析
变压器
人机交互
生态学
物理
管理
量子力学
电压
经济
生物
作者
Jingping Liu,Tao Chen,Zujie Liang,Haiyun Jiang,Yanghua Xiao,Wei Feng,Yuxi Qian,Zhenghong Hao,Bing Han
标识
DOI:10.1145/3583780.3614913
摘要
Prompt tuning has enhanced the performance of Pre-trained Language Models for multi-task learning in few-shot scenarios. However, existing studies fail to consider that the prompts among different layers in Transformer are different due to the diverse information learned at each layer. In general, the bottom layers in the model tend to capture low-level semantic or structural information, while the upper layers primarily acquire task-specific knowledge. Hence, we propose a novel hierarchical prompt tuning model for few-shot multi-task learning to capture this regularity. The designed model mainly consists of three types of prompts: shared prompts, auto-adaptive prompts, and task-specific prompts. Shared prompts facilitate the sharing of general information across all tasks. Auto-adaptive prompts dynamically select and integrate relevant prompt information from all tasks into the current task. Task-specific prompts concentrate on learning task-specific knowledge. To enhance the model's adaptability to diverse inputs, we introduce deep instance-aware language prompts as the foundation for constructing the above prompts. To evaluate the effectiveness of our proposed method, we conduct extensive experiments on multiple widely-used datasets. The experimental results demonstrate that the proposed method achieves state-of-the-art performance for multi-task learning in few-shot settings and outperforms ChatGPT in the full-data setting.
科研通智能强力驱动
Strongly Powered by AbleSci AI