计算机科学
翻译(生物学)
人工智能
图像翻译
面子(社会学概念)
棱锥(几何)
图像(数学)
任务(项目管理)
计算机视觉
过程(计算)
保险丝(电气)
模式识别(心理学)
领域(数学分析)
化学
经济
管理
社会学
工程类
数学分析
物理
光学
电气工程
操作系统
信使核糖核酸
基因
生物化学
社会科学
数学
作者
Changcheng Liang,Mingrui Zhu,Nannan Wang,Heng Yang,Xinbo Gao
标识
DOI:10.1109/tnnls.2022.3233025
摘要
In this article, we address the face image translation task, which aims to translate a face image of a source domain to a target domain. Although significant progress has been made by recent studies, face image translation is still a challenging task because it has more strict requirements for texture details: even a few artifacts will greatly affect the impression of generated face images. Targeting to synthesize high-quality face images with admirable visual appearance, we revisit the coarse-to-fine strategy and propose a novel parallel multistage architecture on the basis of generative adversarial networks (PMSGAN). More specifically, PMSGAN progressively learns the translation function by disintegrating the general synthesis process into multiple parallel stages that take images with gradually decreasing spatial resolution as inputs. To prompt the information exchange between various stages, a cross-stage atrous spatial pyramid (CSASP) structure is specially designed to receive and fuse the contextual information from other stages. At the end of the parallel model, we introduce a novel attention-based module that leverages multistage decoded outputs as in situ supervised attention to refine the final activations and yield the target image. Extensive experiments on several face image translation benchmarks show that PMSGAN performs considerably better than state-of-the-art approaches.
科研通智能强力驱动
Strongly Powered by AbleSci AI