A temporal convolutional recurrent autoencoder based framework for compressing time series data

自编码 循环神经网络 计算机科学 深度学习 卷积神经网络 人工智能 编码器 时间序列 系列(地层学) 数据压缩 模式识别(心理学) 人工神经网络 算法 机器学习 生物 操作系统 古生物学
作者
Zhong Zheng,Zijun Zhang
出处
期刊:Applied Soft Computing [Elsevier BV]
卷期号:147: 110797-110797 被引量:10
标识
DOI:10.1016/j.asoc.2023.110797
摘要

The sharply growing volume of time series data due to recent sensing technology advancement poses emerging challenges to the data transfer speed and storage as well as corresponding energy consumption. To tackle the overwhelming volume of time series data in transmission and storage, compressing time series, which encodes time series into smaller size representations while enables authentic restoration of compressed ones with minimizing the reconstruction error, has attracted significant attention. Numerous methods have been developed and recent deep learning ones with minimal assumptions on data characteristics, such as recurrent autoencoders, have shown themselves to be competitive. Yet, capturing long-term dependencies in time series compression is a significant challenge calling further development. To make a response, this paper proposes a temporal convolutional recurrent autoencoder framework for more effective time series compression. First, two autoencoder modules, the temporal convolutional network encoder with a recurrent neural network decoder (TCN-RNN) and the temporal convolutional network encoder with an attention assisted recurrent neural network decoder (TCN-ARNN), are developed. The TCN-RNN employs only the recurrent neural network decoder to reconstruct the time series in reverse order. In contrast, the TCN-ARNN uses two recurrent neural networks to reconstruct the time series in both forward and reverse order in parallel. In addition, a timestep-wise attention network is developed to incorporate the forward and reverse reconstructions into the ultimate reconstruction with adaptive weights. Finally, a model selection procedure is developed to adaptively select between the TCN-RNN and TCN-ARNN based on their reconstruction performance on the validation dataset. Computational experiments on five datasets show that the proposed temporal convolutional recurrent autoencoder outperforms state-of-the-art benchmarking models in terms of lower reconstruction errors with the same compression ratio, achieving an improvement of up to 45.14% in the average of mean squared errors. Results indicate a promising potential of the proposed temporal convolutional recurrent autoencoder on the time series compression for various applications involving long time series data.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
canghong完成签到,获得积分10
刚刚
风子发布了新的文献求助10
1秒前
1秒前
Sid发布了新的文献求助10
1秒前
1秒前
妮妮完成签到,获得积分10
2秒前
3秒前
3秒前
我是老大应助WY采纳,获得10
4秒前
白术发布了新的文献求助10
5秒前
dde应助JJ采纳,获得10
5秒前
尊敬的千愁完成签到,获得积分10
6秒前
xingchentu完成签到 ,获得积分10
6秒前
6秒前
zhong完成签到 ,获得积分10
8秒前
ttssooe发布了新的文献求助10
10秒前
red发布了新的文献求助10
10秒前
10秒前
11秒前
13秒前
彭于晏应助嘉的科研采纳,获得10
13秒前
大个应助Luobing采纳,获得10
13秒前
魔幻的凡双完成签到,获得积分10
14秒前
14秒前
15秒前
15秒前
16秒前
楠枫完成签到,获得积分10
17秒前
上官若男应助蓝天采纳,获得10
18秒前
李光发布了新的文献求助10
18秒前
18秒前
19秒前
19秒前
酷波er应助科研通管家采纳,获得10
19秒前
Ava应助科研通管家采纳,获得10
19秒前
赘婿应助科研通管家采纳,获得10
19秒前
人间以上发布了新的文献求助30
19秒前
小蘑菇应助科研通管家采纳,获得10
19秒前
CodeCraft应助科研通管家采纳,获得50
19秒前
华仔应助科研通管家采纳,获得10
19秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
PowerCascade: A Synthetic Dataset for Cascading Failure Analysis in Power Systems 2000
Various Faces of Animal Metaphor in English and Polish 800
Fundamentals of Strain Psychology 800
The SAGE Dictionary of Qualitative Inquiry 610
Signals, Systems, and Signal Processing 610
An Introduction to Medicinal Chemistry 第六版习题答案 600
热门求助领域 (近24小时)
化学 材料科学 医学 生物 纳米技术 工程类 有机化学 化学工程 生物化学 计算机科学 物理 内科学 复合材料 催化作用 物理化学 光电子学 电极 细胞生物学 基因 无机化学
热门帖子
关注 科研通微信公众号,转发送积分 6343535
求助须知:如何正确求助?哪些是违规求助? 8158533
关于积分的说明 17152530
捐赠科研通 5399889
什么是DOI,文献DOI怎么找? 2860062
邀请新用户注册赠送积分活动 1838111
关于科研通互助平台的介绍 1687782