计算机科学
人工智能
词汇分析
安全性令牌
特征提取
变压器
语义学(计算机科学)
模式识别(心理学)
程序设计语言
计算机安全
量子力学
物理
电压
作者
Ray Ming,Na Chen,Jiangtao Peng,Weiwei Sun,Zhijing Ye
标识
DOI:10.1109/jstars.2025.3528122
摘要
Recently, the Transformer-based model has shown superior performance in hyperspectral image classification (HSIC) due to its excellent ability to model long-term dependencies on sequence data. An important component of the Transformer is the tokenizer, which can transform the features into semantic token sequences (STS). Nonetheless, Transformer's semantic tokenization strategy is hardly representative of local relatively important high-level semantics because of its global receptive field. Recently, the Mamba-based methods have shown even stronger spatial context modeling ability than Transformer for HSIC. However, these Mamba-based methods mainly focus on spectral and spatial dimensions. They tend to extract semantic information in very long feature sequences or represent semantic information in several typical tokens, which may ignore some important semantics of the HSIs. In order to represent the semantic information of HSIs more holistically in Mamba, this article proposes a semantic tokenization-based Mamba (STMamba) model. In STMamba, a spectral-spatial feature extraction (SSFE) module is used to extract the spectral-spatial joint features. Then a generated semantic token sequences module (GSTSM) is designed to transform the features into STS. Subsequently, the STS are fed into the semantic token state spatial model (STSSM) to capture relationships between different semantic tokens. Finally, the fused semantic token is passed into a classifier for classification. Experimental results on three HSI datasets demonstrate that the proposed STMamba outperforms existing state-of-the-art deep learning and transformer-based methods. The source code can be found at https://github.com/AlanLowell/STMamba.
科研通智能强力驱动
Strongly Powered by AbleSci AI