Skymath: Technical Report · The Large Language Model Bible Contribute to LLM-Bible

Skymath: Technical Report

Yang Liu, Yang Haihua, Cheng Wenjun, Lin Lei, Li Chenxia, Chen Yifu, Liu Lunan, Pan Jianfei, Wei Tianwen, Li Biye, Zhao Liang, Wang Lijie, Zhu Bo, Li Guoliang, Wu Xuejie, Luo Xilin, Hu Rui. Arxiv 2023

[Paper]    
Fine Tuning Pretraining Methods Training Techniques

Large language models (LLMs) have shown great potential to solve varieties of natural language processing (NLP) tasks, including mathematical reasoning. In this work, we present SkyMath, a large language model for mathematics with 13 billion parameters. By applying self-compare fine-tuning, we have enhanced mathematical reasoning abilities of Skywork-13B-Base remarkably. On GSM8K, SkyMath outperforms all known open-source models of similar size and has established a new SOTA performance.

Similar Work