人工智能
先验概率
生成语法
代表(政治)
计算机科学
模式识别(心理学)
生成模型
图像(数学)
特征(语言学)
发电机(电路理论)
编码器
计算机视觉
光学(聚焦)
贝叶斯概率
物理
哲学
光学
操作系统
功率(物理)
政治
法学
量子力学
语言学
政治学
作者
Geonung Kim,Kyoungkook Kang,Seong‐Tae Kim,Hwayoon Lee,Sehoon Kim,Jonghyun Kim,Seung‐Hwan Baek,Sunghyun Cho
出处
期刊:Cornell University - arXiv
日期:2022-01-01
标识
DOI:10.48550/arxiv.2207.09685
摘要
For realistic and vivid colorization, generative priors have recently been exploited. However, such generative priors often fail for in-the-wild complex images due to their limited representation space. In this paper, we propose BigColor, a novel colorization approach that provides vivid colorization for diverse in-the-wild images with complex structures. While previous generative priors are trained to synthesize both image structures and colors, we learn a generative color prior to focus on color synthesis given the spatial structure of an image. In this way, we reduce the burden of synthesizing image structures from the generative prior and expand its representation space to cover diverse images. To this end, we propose a BigGAN-inspired encoder-generator network that uses a spatial feature map instead of a spatially-flattened BigGAN latent code, resulting in an enlarged representation space. Our method enables robust colorization for diverse inputs in a single forward pass, supports arbitrary input resolutions, and provides multi-modal colorization results. We demonstrate that BigColor significantly outperforms existing methods especially on in-the-wild images with complex structures.
科研通智能强力驱动
Strongly Powered by AbleSci AI