计算机科学
水准点(测量)
学习迁移
人工智能
编码(集合论)
特征(语言学)
任务(项目管理)
特征提取
过程(计算)
模式识别(心理学)
机器学习
操作系统
哲学
大地测量学
经济
集合(抽象数据类型)
语言学
管理
程序设计语言
地理
作者
Zhengxu Yu,Dong Shen,Zhongming Jin,Jianqiang Huang,Deng Cai,Xian–Sheng Hua
出处
期刊:IEEE transactions on image processing
[Institute of Electrical and Electronics Engineers]
日期:2022-01-01
卷期号:31: 1340-1348
被引量:11
标识
DOI:10.1109/tip.2022.3141258
摘要
Model fine-tuning is a widely used transfer learning approach in person Re-identification (ReID) applications, which fine-tuning a pre-trained feature extraction model into the target scenario instead of training a model from scratch. It is challenging due to the significant variations inside the target scenario, e.g., different camera viewpoint, illumination changes, and occlusion. These variations result in a gap between each mini-batch's distribution and the whole dataset's distribution when using mini-batch training. In this paper, we study model fine-tuning from the perspective of the aggregation and utilization of the dataset's global information when using mini-batch training. Specifically, we introduce a novel network structure called Batch-related Convolutional Cell (BConv-Cell), which progressively collects the dataset's global information into a latent state and uses it to rectify the extracted feature. Based on BConv-Cells, we further proposed the Progressive Transfer Learning (PTL) method to facilitate the model fine-tuning process by jointly optimizing BConv-Cells and the pre-trained ReID model. Empirical experiments show that our proposal can greatly improve the ReID model's performance on MSMT17, Market-1501, CUHK03, and DukeMTMC-reID datasets. Moreover, we extend our proposal to the general image classification task. The experiments in several image classification benchmark datasets demonstrate that our proposal can significantly improve baseline models' performance. The code has been released at https://github.com/ZJULearning/PTL.
科研通智能强力驱动
Strongly Powered by AbleSci AI