计算机科学
信息隐藏
利用
辍学(神经网络)
编码
可微函数
计算机视觉
深层神经网络
人工神经网络
JPEG格式
编码器
深度学习
人工智能
模式识别(心理学)
机器学习
图像(数学)
基因
操作系统
数学分析
生物化学
化学
计算机安全
数学
作者
Jiren Zhu,Russell Kaplan,Justin Johnson,Li Fei-Fei
标识
DOI:10.1007/978-3-030-01267-0_40
摘要
Recent work has shown that deep neural networks are highly sensitive to tiny perturbations of input images, giving rise to adversarial examples. Though this property is usually considered a weakness of learned models, we explore whether it can be beneficial. We find that neural networks can learn to use invisible perturbations to encode a rich amount of useful information. In fact, one can exploit this capability for the task of data hiding. We jointly train encoder and decoder networks, where given an input message and cover image, the encoder produces a visually indistinguishable encoded image, from which the decoder can recover the original message. We show that these encodings are competitive with existing data hiding algorithms, and further that they can be made robust to noise: our models learn to reconstruct hidden information in an encoded image despite the presence of Gaussian blurring, pixel-wise dropout, cropping, and JPEG compression. Even though JPEG is non-differentiable, we show that a robust model can be trained using differentiable approximations. Finally, we demonstrate that adversarial training improves the visual quality of encoded images.
科研通智能强力驱动
Strongly Powered by AbleSci AI