航空
计算机科学
航空安全
航空事故
领域(数学分析)
变压器
民用航空
叙述的
工程类
数学分析
语言学
哲学
数学
电气工程
电压
航空航天工程
作者
Chetan Chandra,Jiantao Xiao,Mayank V. Bendarkar,Kshitij Sawant,Lidya Elias,Michelle Kirby,Dimitri N. Mavris
摘要
Data-driven methods form the frontier of reactive aviation safety analysis. While analysis of quantitative data from flight operations is common, text narratives of accidents and incidents have not been sufficiently mined. Among the many use cases of aviation text-data mining, automatically extracting safety concepts is probably the most important. Bidirectional Encoder Representations from Transformers (BERT) is a transformer-based large language model that is openly available and has been adapted to numerous domain-specific tasks. The present work provides a comprehensive methodology to develop domain-specific BERT model starting from the base model. A preliminary aviation domain-specific BERT model is developed in this work. This Aviation-BERT model is pre-trained from the BERT-Base model using accident and incident text narratives from the National Transportation Safety Board (NTSB) and Aviation Safety Reporting System (ASRS) using mixed-domain pre-training. Aviation-BERT is shown to outperform BERT when it comes to text-mining tasks on aviation text datasets. It is also expected to be of tremendous value in numerous downstream tasks in the analysis of aviation text corpora.
科研通智能强力驱动
Strongly Powered by AbleSci AI