图像扭曲
人工智能
图像翻译
计算机科学
翻译(生物学)
图像(数学)
计算机视觉
模式识别(心理学)
失真(音乐)
生成语法
信使核糖核酸
基因
生物化学
化学
放大器
带宽(计算)
计算机网络
作者
Lin-Chieh Huang,Hung-Hsu Tsai
标识
DOI:10.1016/j.neunet.2023.07.010
摘要
This paper proposes an unsupervised image-to-image (UI2I) translation model, called Perceptual Contrastive Generative Adversarial Network (PCGAN), which can mitigate the distortion problem to enhance performance of the traditional UI2I methods. The PCGAN is designed with a two-stage UI2I model. In the first stage of the PCGAN, it leverages a novel image warping to transform shapes of objects in input (source) images. In the second stage of the PCGAN, the residual prediction is devised in refinements of the outputs of the first stage of the PCGAN. To promote performance of the image warping, a loss function, called Perceptual Patch-Wise InfoNCE, is developed in the PCGAN to effectively memorize the visual correspondences between warped images and refined images. Experimental results on quantitative evaluation and visualization comparison for UI2I benchmarks show that the PCGAN is superior to other existing methods considered here.
科研通智能强力驱动
Strongly Powered by AbleSci AI