Mutual Enhancement Of Large And Small Language Models With Cross-silo Knowledge Transfer · The Large Language Model Bible Contribute to LLM-Bible

Mutual Enhancement Of Large And Small Language Models With Cross-silo Knowledge Transfer

Deng Yongheng, Qiao Ziqing, Ren Ju, Liu Yang, Zhang Yaoxue. Arxiv 2023

[Paper]    
Fine Tuning Pretraining Methods Training Techniques

While large language models (LLMs) are empowered with broad knowledge, their task-specific performance is often suboptimal. It necessitates fine-tuning LLMs with task-specific data, but such data may be inaccessible due to privacy concerns. In this paper, we propose a novel approach to enhance LLMs with smaller language models (SLMs) that are trained on clients using their private task-specific data. To enable mutual enhancement between LLMs and SLMs, we propose CrossLM, where the SLMs promote the LLM to generate task-specific high-quality data, and both the LLM and SLMs are enhanced with the generated data. We evaluate CrossLM using publicly accessible language models across a range of benchmark tasks. The results demonstrate that CrossLM significantly enhances the task-specific performance of SLMs on clients and the LLM on the cloud server simultaneously while preserving the LLM’s generalization capability.

Similar Work