离散余弦变换
卷积(计算机科学)
核(代数)
卷积定理
快速傅里叶变换
卷积神经网络
计算机科学
算法
圆卷积
重叠-添加方法
人工智能
离散傅里叶变换(通用)
频域
数学
特征(语言学)
傅里叶变换
模式识别(心理学)
人工神经网络
分数阶傅立叶变换
计算机视觉
离散数学
图像(数学)
数学分析
傅里叶分析
哲学
语言学
作者
Yuhao Xu,Hideki Nakayama
标识
DOI:10.1109/ijcnn52387.2021.9534135
摘要
Spectral representations have been introduced into deep convolutional neural networks (CNNs) mainly for accelerating convolutions and mitigating information loss. However, repeated domain transformations and complex arithmetic of commonly-used Fourier transform (DFT, FFT) seriously limit the applicability of spectral networks. In contrast, discrete cosine transform (DCT)-based methods are more promising owing to computations with merely real numbers. Hence in this work, we investigate the convolution theorem of DCT and propose a faster spectral convolution method for CNNs. First, we transform the input feature map and convolutional kernel into the frequency domain via DCT. We then perform element-wise multiplication between the spectral feature map and kernel, which is mathematically equivalent to symmetric convolution in the spatial domain but much cheaper than the straightforward spatial convolution. Since DCT only involves real arithmetic, the computational complexity of our method is significantly smaller than the traditional FFT-based spectral convolution. Besides, we introduce a network optimization strategy to suppress repeated domain transformations leveraging the intrinsically extended kernels. Furthermore, we present a partial symmetry breaking strategy with spectral dropout to mitigate the performance degradation caused by kernel symmetry. Experimental results demonstrate that compared with traditional spatial and spectral methods, our proposed DCT-based spectral convolution effectively accelerates the networks while achieving comparable accuracy.
科研通智能强力驱动
Strongly Powered by AbleSci AI