变压器
计算机科学
编码(集合论)
人工智能
自然语言处理
一般化
生成语法
质量(理念)
建筑
中医药
医学
替代医学
程序设计语言
病理
艺术
数学分析
哲学
物理
数学
集合(抽象数据类型)
认识论
量子力学
电压
视觉艺术
作者
Yang Tan,Yang Tan,Mingchen Li,Fei Pan,Hao Duan,Zijie Huang,Hua Deng,Zhuohang Yu,Chen Yang,Guoyang Shen,Peng Qi,Chengyuan Yue,Yuxian Liu,Liang Hong,Huiqun Yu,Guisheng Fan,Yun Tang
标识
DOI:10.1016/j.compbiomed.2024.108290
摘要
Generative Large Language Models (LLMs) have achieved significant success in various natural language processing tasks, including Question-Answering (QA) and dialogue systems. However, most models are trained on English data and lack strong generalization in providing answers in Chinese. This limitation is especially evident in specialized domains like traditional Chinese medical QA, where performance suffers due to the absence of fine-tuning and high-quality datasets. To address this, we introduce MedChatZH, a dialogue model optimized for Chinese medical QA based on transformer decoder with LLaMA architecture. Continued pre-training on a curated corpus of Chinese medical books is followed by fine-tuning with a carefully selected medical instruction dataset, resulting in MedChatZH outperforming several Chinese dialogue baselines on a real-world medical dialogue dataset. Our model, code, and dataset are publicly available on GitHub (https://github.com/tyang816/MedChatZH) to encourage further research in traditional Chinese medicine and LLMs.
科研通智能强力驱动
Strongly Powered by AbleSci AI