计算机科学
基线(sea)
计算复杂性理论
计算模型
GSM演进的增强数据速率
特征提取
深度学习
语音识别
功率(物理)
边缘设备
机器学习
人工智能
算法
量子力学
云计算
海洋学
操作系统
物理
地质学
作者
Xiaobin Rong,Tianchi Sun,Xu Zhang,Yuxiang Hu,Changbao Zhu,Jing Lü
标识
DOI:10.1109/icassp48485.2024.10448310
摘要
While modern deep learning-based models have significantly outperformed traditional methods in the area of speech enhancement, they often necessitate a lot of parameters and extensive computational power, making them impractical to be deployed on edge devices in real-world applications. In this paper, we introduce Grouped Temporal Convolutional Recurrent Network (GTCRN), which incorporates grouped strategies to efficiently simplify a competitive model, DPCRN. Additionally, it leverages subband feature extraction modules and temporal recurrent attention modules to enhance its performance. Remarkably, the resulting model demands ultralow computational resources, featuring only 23.7 K parameters and 39.6 MMACs per second. Experimental results show that our proposed model not only surpasses RNNoise, a typical lightweight model with similar computational burden, but also achieves competitive performance when compared to recent baseline models with significantly higher computational resources requirements.
科研通智能强力驱动
Strongly Powered by AbleSci AI