生物
适配器(计算)
杠杆(统计)
计算生物学
人工智能
机器学习
计算机科学
操作系统
作者
Shuai Zeng,Duolin Wang,Lei Jiang,Dong Xu
出处
期刊:Genome Research
[Cold Spring Harbor Laboratory]
日期:2024-07-26
卷期号:: gr.279132.124-gr.279132.124
被引量:2
标识
DOI:10.1101/gr.279132.124
摘要
Signal peptides (SP) play a crucial role in protein translocation in cells. The development of large protein language models (PLMs) and prompt-based learning provides a new opportunity for SP prediction, especially for the categories with limited annotated data. We present a parameter-efficient fine-tuning (PEFT) framework for SP prediction, PEFT-SP, to effectively utilize pretrained PLMs. We integrated low-rank adaptation (LoRA) into ESM-2 models to better leverage the protein sequence evolutionary knowledge of PLMs. Experiments show that PEFT-SP using LoRA enhances state-of-the-art results, leading to a maximum Matthews correlation coefficient (MCC) gain of 87.3% for SPs with small training samples and an overall MCC gain of 6.1%. Furthermore, we also employed two other PEFT methods, prompt tuning and adapter tuning, in ESM-2 for SP prediction. More elaborate experiments show that PEFT-SP using adapter tuning can also improve the state-of-the-art results by up to 28.1% MCC gain for SPs with small training samples and an overall MCC gain of 3.8%. LoRA requires fewer computing resources and less memory than the adapter during the training stage, making it possible to adapt larger and more powerful protein models for SP prediction.
科研通智能强力驱动
Strongly Powered by AbleSci AI