计算机科学
任务(项目管理)
集合(抽象数据类型)
模式(计算机接口)
自然语言处理
人工智能
序列(生物学)
训练集
机器翻译
语言模型
语音识别
机器学习
程序设计语言
人机交互
遗传学
生物
经济
管理
作者
Chang Su,Xiaofeng Zhao,Xiaosong Qiao,Min Zhang,Hao Yang,Junhao Zhu,Ming Zhu,Wenbin Ma
标识
DOI:10.1007/978-3-031-44699-3_6
摘要
Deep learning has shown remarkable effectiveness in various language tasks. This paper presents Huawei Translation Services Center’s (HW-TSC’s) work called HWCGEC which get the best performance among the seven submitted results in the NLPCC2023 shared task 1, namely Chinese grammatical error correction (CGEC). CGEC aims to automatically correct grammatical errors that violate language rules and converts the noisy input texts to clean output texts. This paper, through experiments, discovered that after model fine-tuning the BART a sequence to sequence (seq2seq) model performs better than the ChatGLM a large language model (LLM) in situations where training data is large while the LoRA mode has a smaller number of parameters for fine-tuning. Additionally, the BART model achieves good results in the CGEC task through data augmentation and curriculum learning methods. Although the performance of LLM is poor in experiments, they possess excellent logical abilities. With the training set becoming more diverse and the methods for training set data augmentation becoming more refined, the supervised fine-tuning (SFT) mode trained LLMs are expected to achieve significant improvements in CGEC tasks in the future.
科研通智能强力驱动
Strongly Powered by AbleSci AI