计算机科学
人工智能
自然语言处理
稳健性(进化)
机器学习
语言模型
深度学习
财务
构造(python库)
训练集
生物化学
基因
经济
化学
程序设计语言
作者
Zhuang Liu,Degen Huang,Kaiyu Huang,Zhuang Li,Jun Zhao
标识
DOI:10.24963/ijcai.2020/622
摘要
There is growing interest in the tasks of financial text mining. Over the past few years, the progress of Natural Language Processing (NLP) based on deep learning advanced rapidly. Significant progress has been made with deep learning showing promising results on financial text mining models. However, as NLP models require large amounts of labeled training data, applying deep learning to financial text mining is often unsuccessful due to the lack of labeled training data in financial fields. To address this issue, we present FinBERT (BERT for Financial Text Mining) that is a domain specific language model pre-trained on large-scale financial corpora. In FinBERT, different from BERT, we construct six pre-training tasks covering more knowledge, simultaneously trained on general corpora and financial domain corpora, which can enable FinBERT model better to capture language knowledge and semantic information. The results show that our FinBERT outperforms all current state-of-the-art models. Extensive experimental results demonstrate the effectiveness and robustness of FinBERT. The source code and pre-trained models of FinBERT are available online.
科研通智能强力驱动
Strongly Powered by AbleSci AI