人工智能
计算机科学
可解释性
自编码
模式识别(心理学)
机器学习
深度学习
上下文图像分类
特征学习
特征(语言学)
遮罩(插图)
特征提取
图像(数学)
艺术
语言学
哲学
视觉艺术
作者
Zong Fan,Zhimin Wang,Ping Gong,Christine U. Lee,Shanshan Tang,Xiaohui Zhang,Yao Hao,Zhongwei Zhang,Pengfei Song,Shigao Chen,Li Hua
摘要
Accurate classification of medical images is crucial for disease diagnosis and treatment planning. Deep learning (DL) methods have gained increasing attention in this domain. However, DL-based classification methods encounter challenges due to the unique characteristics of medical image datasets, including limited amounts of labeled images and large image variations. Self-supervised learning (SSL) has emerged as a solution that learns informative representations from unlabeled data to alleviate the scarcity of labeled images and improve model performance. A recently proposed generative SSL method, masked autoencoder (MAE), has shown excellent capability in feature representation learning. The MAE model trained on unlabeled data can be easily tuned to improve the performance of various downstream classification models. In this paper, we performed a preliminary study to integrate MAE with the self-attention mechanism for tumor classification on breast ultrasound (BUS) data. Considering the speckle noise, image quality variations of BUS images, and varying tumor shapes and sizes, two revisions were adopted in using MAE for tumor classification. First, MAE's patch size and masking ratio were adjusted to avoid missing information embedded in small lesions on BUS images. Second, attention maps were extracted to improve the interpretability of the model's decision-making process. Experiments demonstrated the effectiveness and potential of the MAE-based classification model on small labeled datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI