计算机科学
人工智能
自然语言处理
计算生物学
生物
作者
Yucheng Xia,Yuhang Liu,Tianhao Li,Sihan He,Hong Chang,Yaqing Wang,Yongqing Zhang,Wenyi Ge
出处
期刊:Methods
[Elsevier]
日期:2024-05-15
卷期号:228: 12-21
被引量:1
标识
DOI:10.1016/j.ymeth.2024.05.007
摘要
Annotating cell types of single-cell RNA sequencing (scRNA-seq) data is crucial for studying cellular heterogeneity in the tumor microenvironment. Recently, large-scale pre-trained language models (PLMs) have achieved significant progress in cell-type annotation of scRNA-seq data. This approach effectively addresses previous methods' shortcomings in performance and generalization. However, fine-tuning PLMs for different downstream tasks demands considerable computational resources, rendering it impractical. Hence, a new research branch introduces parameter-efficient fine-tuning (PEFT). This involves optimizing a few parameters while leaving the majority unchanged, leading to substantial reductions in computational expenses. Here, we utilize scBERT, a large-scale pre-trained model, to explore the capabilities of three PEFT methods in scRNA-seq cell type annotation. Extensive benchmark studies across several datasets demonstrate the superior applicability of PEFT methods. Furthermore, downstream analysis using models obtained through PEFT showcases their utility in novel cell type discovery and model interpretability for potential marker genes. Our findings underscore the considerable potential of PEFT in PLM-based cell type annotation, presenting novel perspectives for the analysis of scRNA-seq data.
科研通智能强力驱动
Strongly Powered by AbleSci AI