变压器
计算机科学
融合
模式治疗法
调制(音乐)
语音识别
人工智能
工程类
电气工程
医学
物理
电压
哲学
语言学
外科
声学
作者
Menɡ Ninɡ,Fan Zhou,Wei Wang,Shaoqiang Wang,Peiying Zhang,Jian Wang
出处
期刊:Electronics
[MDPI AG]
日期:2024-09-20
卷期号:13 (18): 3725-3725
标识
DOI:10.3390/electronics13183725
摘要
Multimodal automatic modulation recognition (MAMR) has emerged as a prominent research area. The effective fusion of features from different modalities is crucial for MAMR tasks. An effective multimodal fusion mechanism should maximize the extraction and integration of complementary information. Recently, fusion methods based on cross-modal attention have shown high performance. However, they overlook the differences in information intensity between different modalities, suffering from quadratic complexity. To this end, we propose an efficient Alignment before Fusion Transformer Network (AbFTNet) based on an in-phase quadrature (I/Q) and Fractional Fourier Transform (FRFT). Specifically, we first align and correlate the feature representations of different single modalities to achieve mutual information maximization. The single modality feature representations are obtained using the self-attention mechanism of the Transformer. Then, we design an efficient cross-modal aggregation promoting (CAP) module. By designing the aggregation center, we integrate two modalities to achieve the adaptive complementary learning of modal features. This operation bridges the gap in information intensity between different modalities, enabling fair interaction. To verify the effectiveness of the proposed methods, we conduct experiments on the RML2016.10a dataset. The experimental results show that multimodal fusion features significantly outperform single-modal features in classification accuracy across different signal-to-noise ratios (SNRs). Compared to other methods, AbFTNet achieves an average accuracy of 64.59%, with a 1.36% improvement over the TLDNN method, reaching the state of the art.
科研通智能强力驱动
Strongly Powered by AbleSci AI