循环神经网络
编码器
计算机科学
变压器
卷积神经网络
深度学习
人工智能
短时记忆
序列学习
自编码
模式识别(心理学)
算法
人工神经网络
工程类
电压
电气工程
操作系统
作者
Ziao Yang,Xiangrui Yang,Qifeng Lin
出处
期刊:Cornell University - arXiv
日期:2021-12-02
被引量:1
摘要
Spatiotemporal predictive learning is to generate future frames given a sequence of historical frames. Conventional algorithms are mostly based on recurrent neural networks (RNNs). However, RNN suffers from heavy computational burden such as time and long back-propagation process due to the seriality of recurrent structure. Recently, Transformer-based methods have also been investigated in the form of encoder-decoder or plain encoder, but the encoder-decoder form requires too deep networks and the plain encoder is lack of short-term dependencies. To tackle these problems, we propose an algorithm named 3D-temporal convolutional transformer (TCTN), where a transformer-based encoder with temporal convolutional layers is employed to capture short-term and long-term dependencies. Our proposed algorithm can be easy to implement and trained much faster compared with RNN-based methods thanks to the parallel mechanism of Transformer. To validate our algorithm, we conduct experiments on the MovingMNIST and KTH dataset, and show that TCTN outperforms state-of-the-art (SOTA) methods in both performance and training speed.
科研通智能强力驱动
Strongly Powered by AbleSci AI