计算机科学
变压器
分割
人工智能
特征学习
机器学习
模式识别(心理学)
工程类
电压
电气工程
作者
Pei He,Licheng Jiao,Ronghua Shang,Xu Liu,Fang Liu,Shuyuan Yang,Xiangrong Zhang,Shuang Wang
出处
期刊:IEEE transactions on neural networks and learning systems
[Institute of Electrical and Electronics Engineers]
日期:2023-06-07
卷期号:35 (10): 14138-14150
被引量:5
标识
DOI:10.1109/tnnls.2023.3274760
摘要
Domain generalization (DG) is one of the critical issues for deep learning in unknown domains. How to effectively represent domain-invariant context (DIC) is a difficult problem that DG needs to solve. Transformers have shown the potential to learn generalized features, since the powerful ability to learn global context. In this article, a novel method named patch diversity Transformer (PDTrans) is proposed to improve the DG for scene segmentation by learning global multidomain semantic relations. Specifically, patch photometric perturbation (PPP) is proposed to improve the representation of multidomain in the global context information, which helps the Transformer learn the relationship between multiple domains. Besides, patch statistics perturbation (PSP) is proposed to model the feature statistics of patches under different domain shifts, which enables the model to encode domain-invariant semantic features and improve generalization. PPP and PSP can help to diversify the source domain at the patch level and feature level. PDTrans learns context across diverse patches and takes advantage of self-attention to improve DG. Extensive experiments demonstrate the tremendous performance advantages of the PDTrans over state-of-the-art DG methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI