A temporal convolutional recurrent autoencoder based framework for compressing time series data

自编码 循环神经网络 计算机科学 深度学习 卷积神经网络 人工智能 编码器 时间序列 系列(地层学) 数据压缩 模式识别(心理学) 人工神经网络 算法 机器学习 生物 操作系统 古生物学
作者
Zhong Zheng,Zijun Zhang
出处
期刊:Applied Soft Computing [Elsevier BV]
卷期号:147: 110797-110797 被引量:10
标识
DOI:10.1016/j.asoc.2023.110797
摘要

The sharply growing volume of time series data due to recent sensing technology advancement poses emerging challenges to the data transfer speed and storage as well as corresponding energy consumption. To tackle the overwhelming volume of time series data in transmission and storage, compressing time series, which encodes time series into smaller size representations while enables authentic restoration of compressed ones with minimizing the reconstruction error, has attracted significant attention. Numerous methods have been developed and recent deep learning ones with minimal assumptions on data characteristics, such as recurrent autoencoders, have shown themselves to be competitive. Yet, capturing long-term dependencies in time series compression is a significant challenge calling further development. To make a response, this paper proposes a temporal convolutional recurrent autoencoder framework for more effective time series compression. First, two autoencoder modules, the temporal convolutional network encoder with a recurrent neural network decoder (TCN-RNN) and the temporal convolutional network encoder with an attention assisted recurrent neural network decoder (TCN-ARNN), are developed. The TCN-RNN employs only the recurrent neural network decoder to reconstruct the time series in reverse order. In contrast, the TCN-ARNN uses two recurrent neural networks to reconstruct the time series in both forward and reverse order in parallel. In addition, a timestep-wise attention network is developed to incorporate the forward and reverse reconstructions into the ultimate reconstruction with adaptive weights. Finally, a model selection procedure is developed to adaptively select between the TCN-RNN and TCN-ARNN based on their reconstruction performance on the validation dataset. Computational experiments on five datasets show that the proposed temporal convolutional recurrent autoencoder outperforms state-of-the-art benchmarking models in terms of lower reconstruction errors with the same compression ratio, achieving an improvement of up to 45.14% in the average of mean squared errors. Results indicate a promising potential of the proposed temporal convolutional recurrent autoencoder on the time series compression for various applications involving long time series data.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
哭泣觅儿完成签到,获得积分20
刚刚
刚刚
不期而遇完成签到 ,获得积分10
1秒前
KoamBggOIp完成签到,获得积分10
2秒前
2秒前
anyway发布了新的文献求助10
2秒前
科研通AI2S应助单身的海白采纳,获得10
2秒前
2秒前
hyscoll发布了新的文献求助10
3秒前
6秒前
听话的梦岚完成签到,获得积分10
6秒前
南明完成签到,获得积分10
6秒前
阿石创吃大餐完成签到 ,获得积分10
6秒前
7秒前
noobking_996发布了新的文献求助50
7秒前
活一下吧发布了新的文献求助10
7秒前
10秒前
Biubiubiu发布了新的文献求助10
10秒前
ji发布了新的文献求助10
11秒前
lusgul完成签到 ,获得积分10
12秒前
14秒前
大个应助hhn采纳,获得10
14秒前
激昂的蜗牛完成签到,获得积分10
15秒前
沉默的幻枫完成签到,获得积分10
16秒前
小孟完成签到,获得积分10
16秒前
赘婿应助memorise采纳,获得10
17秒前
称心的大米完成签到,获得积分10
17秒前
徐宇鹏发布了新的文献求助10
17秒前
云002完成签到 ,获得积分10
18秒前
19秒前
Yang_728发布了新的文献求助10
19秒前
大模型应助干净的琦采纳,获得10
19秒前
wangjq完成签到,获得积分20
20秒前
20秒前
wdhh发布了新的文献求助10
20秒前
20秒前
20秒前
qiqi完成签到,获得积分10
21秒前
22秒前
鲜艳的楼房完成签到,获得积分10
23秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
PowerCascade: A Synthetic Dataset for Cascading Failure Analysis in Power Systems 2000
Metallurgy at high pressures and high temperatures 2000
Various Faces of Animal Metaphor in English and Polish 800
An Introduction to Medicinal Chemistry 第六版习题答案 600
Cleopatra : A Reference Guide to Her Life and Works 500
Fundamentals of Strain Psychology 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 纳米技术 工程类 有机化学 化学工程 生物化学 计算机科学 物理 内科学 复合材料 催化作用 物理化学 光电子学 电极 细胞生物学 基因 无机化学
热门帖子
关注 科研通微信公众号,转发送积分 6341309
求助须知:如何正确求助?哪些是违规求助? 8156666
关于积分的说明 17143731
捐赠科研通 5397490
什么是DOI,文献DOI怎么找? 2859245
邀请新用户注册赠送积分活动 1837192
关于科研通互助平台的介绍 1687226