计算机科学
人工智能
图像(数学)
代表(政治)
图像分割
分割
对抗制
计算机视觉
生成语法
模式识别(心理学)
自然语言处理
政治学
政治
法学
作者
Tomoki Haruyama,Ren Togo,Keisuke Maeda,Takahiro Ogawa,Miki Haseyama
标识
DOI:10.1109/icip42928.2021.9506601
摘要
We propose a novel approach that improves text-guided image manipulation performance in this paper. Text-guided image manipulation aims at modifying some parts of an input image in accordance with the user's text description by semantically associating the regions of the image with the text description. We tackle the conventional methods' problem of modifying undesired parts caused by differences in representation ability between text descriptions and images. Humans tend to pay attention primarily to objects corresponding to the foreground of images, and text descriptions by humans mostly represent the foreground. Therefore, it is necessary to introduce not only a foreground-aware bias based on text descriptions but also a background-aware bias that the text descriptions do not represent. We introduce an image segmentation network into the generative adversarial network for image manipulation to solve the above problem. Comparative experiments with three state-of-the-art methods show the effectiveness of our method quantitatively and qualitatively.
科研通智能强力驱动
Strongly Powered by AbleSci AI