Junyu Zhu,Xu Liao,Changshi Li,Yuling Jiao,Jin Liu,Xiliang Lu
标识
DOI:10.1109/ijcnn54540.2023.10191460
摘要
Improving the generalization of neural networks under domain shift is an important and challenging task in computer vision. Obtaining an invariant representation across domains is a benchmark method in the literature. In this paper, we propose an invariant and sufficient supervised representation learning (ISSRL) approach to learn a domain invariant representation which is also preserving information used for downstream tasks. To this end, we formulate ISSRL by finding a nonlinear map $\boldsymbol{g}$ such that $Y\perp X\vert \boldsymbol{g}(X)$ and $(Y,\boldsymbol{g}(X))\perp D$ at the population level, where D is the label of the domains and $(X, Y)$ is the paired data sampled from domains with label. We use distance correlation to characterize the (conditional) independence. At the sample level, we construct a novel loss function through an unbiased empirical version of distance correlation. We train the representation map by parameterizing it with deep neural networks. Both simulation study and real data evaluation show that ISSRL outperforms the state-of-the-art on out-of-distribution generalization. The PyTorch code for ISSRL is available at https://github.com/CaC033/ISSRL.