计算机科学
语言模型
计算模型
比例(比率)
人工智能
机器学习
理论计算机科学
量子力学
物理
作者
Lei Wang,Hui Zhang,Wei Xu,Zhidong Xue,Yan Wang
标识
DOI:10.1016/j.xcrp.2023.101600
摘要
Protein language models have been gaining attention and have achieved some exciting breakthroughs in protein modeling tasks compared with the utilization of co-evolutionary and biological priors. To overcome the shortcomings of existing large-scale protein language models, such as high computational complexity and large memory consumption, we propose a lightweight protein language model, ProtFlash. It is the first protein language model with linear complexity based on an attention strategy, which differs significantly from existing methods in that it primarily utilizes multiple positional encodings and a mixed chunk attention mechanism that combines local and global attention. Furthermore, the results of the Tasks Assessing Protein Embeddings show that ProtFlash achieves a better or equivalent performance than the state-of-the-art protein language models. As shown through a rigorous comparison of computational complexity and memory consumption, it also performs better than existing protein language models.
科研通智能强力驱动
Strongly Powered by AbleSci AI