计算机科学
概化理论
推论
机器学习
人工智能
稳健性(进化)
图形
特征学习
Boosting(机器学习)
代表(政治)
关系(数据库)
上下文图像分类
潜变量
模式识别(心理学)
数据挖掘
图像(数学)
理论计算机科学
数学
生物化学
统计
化学
政治
政治学
法学
基因
作者
Xian Zhong,Cheng Gu,Mang Ye,Wenxin Huang,Chia‐Wen Lin
标识
DOI:10.1109/tmm.2022.3141886
摘要
Few-shot learning is a tough topic to solve since obtaining a large number of training samples in real applications is challenging. It has attracted increasing attention recently. Meta-learning is a prominent way to address this issue, intending to adapt predictors as base-learners to new tasks swiftly. However, a key challenge of meta-learning is its lack of expressive capacity, which stems from the difficulty of extracting general information from a small number of training samples. As a result, the generalizability of meta-learners trained from high-dimensional parameter spaces is frequently limited. To learn a better representation, we propose a graph complemented latent representation (GCLR) network for few-shot image classification. In particular, we embed the representation into a latent space, in which the latent codes are reconstructed using variational information to enrich the representation. In this way, the latent representation can achieve better generalizability. Another benefit is that, because the latent space is formed using variational inference, it cooperates well with various base-learners, boosting robustness. To make full use of the relation between samples in each category, a graph neural network (GNN) is also incorporated to improve relation mining. Consequently, our end-to-end framework delivers competitive performance on three few-shot learning benchmarks for image classification.
科研通智能强力驱动
Strongly Powered by AbleSci AI