元学习(计算机科学)
计算机科学
软件部署
水准点(测量)
任务(项目管理)
人工智能
机器学习
适应(眼睛)
基线(sea)
弹丸
元建模
软件工程
光学
物理
地质学
经济
有机化学
化学
海洋学
管理
地理
大地测量学
作者
Yong Wu,Shekhor Chanda,Mehrdad Hosseinzadeh,Zhi Liu,Yan Wang
标识
DOI:10.1109/wacv56688.2023.00620
摘要
We consider a new problem of few-shot learning of com-pact models. Meta-learning is a popular approach for few-shot learning. Previous work in meta-learning typically assumes that the model architecture during meta-training is the same as the model architecture used for final deployment. In this paper, we challenge this basic assumption. For final deployment, we often need the model to be small. But small models usually do not have enough capacity to effectively adapt to new tasks. In the mean time, we often have access to the large dataset and extensive computing power during meta-training since meta-training is typically per-formed on a server. In this paper, we propose task-specific meta distillation that simultaneously learns two models in meta-learning: a large teacher model and a small student model. These two models are jointly learned during meta-training. Given a new task during meta-testing, the teacher model is first adapted to this task, then the adapted teacher model is used to guide the adaptation of the student model. The adapted student model is used for final deployment. We demonstrate the effectiveness of our approach in few-shot image classification using model-agnostic meta-learning (MAML). Our proposed method outperforms other alternatives on several benchmark datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI