计算机科学
计算机视觉
光流
人工智能
编码器
运动补偿
运动矢量
运动估计
四分之一像素运动
块匹配算法
规范化(社会学)
运动(物理)
运动插值
视频处理
图像(数学)
视频跟踪
操作系统
社会学
人类学
作者
Jiangning Zhang,Chao Xu,Liang Liu,Mengmeng Wang,Xia Wu,Yong Liu,Yunliang Jiang
标识
DOI:10.1007/978-3-030-58558-7_18
摘要
AbstractThis paper presents a novel end-to-end dynamic time-lapse video generation framework, named DTVNet, to generate diversified time-lapse videos from a single landscape image, which are conditioned on normalized motion vectors. The proposed DTVNet consists of two submodules: Optical Flow Encoder (OFE) and Dynamic Video Generator (DVG). The OFE maps a sequence of optical flow maps to a normalized motion vector that encodes the motion information inside the generated video. The DVG contains motion and content streams that learn from the motion vector and the single image respectively, as well as an encoder and a decoder to learn shared content features and construct video frames with corresponding motion respectively. Specifically, the motion stream introduces multiple adaptive instance normalization (AdaIN) layers to integrate multi-level motion information that are processed by linear layers. In the testing stage, videos with the same content but various motion information can be generated by different normalized motion vectors based on only one input image. We further conduct experiments on Sky Time-lapse dataset, and the results demonstrate the superiority of our approach over the state-of-the-art methods for generating high-quality and dynamic videos, as well as the variety for generating videos with various motion information (https://github.com/zhangzjn/DTVNet).KeywordsGenerative adversarial networkOptical flow encodingTime-Lapse video generation
科研通智能强力驱动
Strongly Powered by AbleSci AI