计算机科学
卷积神经网络
特征(语言学)
人工智能
光学(聚焦)
模式识别(心理学)
特征提取
利用
领域(数学分析)
Mel倒谱
机器学习
语音识别
哲学
数学分析
语言学
数学
物理
计算机安全
光学
作者
Kalpeshkumar Ranipa,Wei‐Ping Zhu,M.N.S. Swamy
标识
DOI:10.1016/j.cmpb.2024.108122
摘要
Most of the existing machine learning-based heart sound classification methods achieve limited accuracy. Since they primarily depend on single domain feature information and tend to focus equally on each part of the signal rather than employing a selective attention mechanism. In addition, they fail to exploit convolutional neural network (CNN) - based features with an effective fusion strategy. In order to overcome these limitations, a novel multimodal attention convolutional neural network (MACNN) with a feature-level fusion strategy, in which Mel-cepstral domain as well as general frequency domain features are incorporated to increase the diversity of the features, is proposed in this paper. In the proposed method, DilationAttenNet is first utilized to construct attention-based CNN feature extractors and then these feature extractors are jointly optimized in MACNN at the feature-level. The attention mechanism aims to suppress irrelevant information and focus on crucial diverse features extracted from the CNN. Extensive experiments are carried out to study the efficacy of the feature level fusion in comparison to that with early fusion. The results show that the proposed MACNN method significantly outperforms the state-of-the-art approaches in terms of accuracy and score for the two publicly available Github and Physionet datasets. The findings of our experiments demonstrated the high performance for heart sound classification based on the proposed MACNN, and hence have potential clinical usefulness in the identification of heart diseases. This technique can assist cardiologists and researchers in the design and development of heart sound classification methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI