Chimed-gpt: A Chinese Medical Large Language Model With Full Training Regime And Better Alignment To Human Preferences · The Large Language Model Bible Contribute to LLM-Bible

Chimed-gpt: A Chinese Medical Large Language Model With Full Training Regime And Better Alignment To Human Preferences

Tian Yuanhe, Gan Ruyi, Song Yan, Zhang Jiaxing, Zhang Yongdong. Arxiv 2023

[Paper] [Code]    
Applications Ethics And Bias Fine Tuning GPT Has Code Model Architecture Pretraining Methods Prompting RAG Reinforcement Learning Training Techniques

Recently, the increasing demand for superior medical services has highlighted the discrepancies in the medical infrastructure. With big data, especially texts, forming the foundation of medical services, there is an exigent need for effective natural language processing (NLP) solutions tailored to the healthcare domain. Conventional approaches leveraging pre-trained models present promising results in this domain and current large language models (LLMs) offer advanced foundation for medical text processing. However, most medical LLMs are trained only with supervised fine-tuning (SFT), even though it efficiently empowers LLMs to understand and respond to medical instructions but is ineffective in learning domain knowledge and aligning with human preference. In this work, we propose ChiMed-GPT, a new benchmark LLM designed explicitly for Chinese medical domain, and undergoes a comprehensive training regime with pre-training, SFT, and RLHF. Evaluations on tasks including information extraction, question answering, and dialogue generation demonstrate ChiMed-GPT’s superior performance over general domain LLMs. Furthermore, we analyze possible biases through prompting ChiMed-GPT to perform attitude scales regarding discrimination of patients, so as to contribute to further responsible development of LLMs in the medical domain. The code and model are released at https://github.com/synlp/ChiMed-GPT.

Similar Work