Preserving Knowledge In Large Language Model With Model-agnostic Self-decompression · The Large Language Model Bible Contribute to LLM-Bible

Preserving Knowledge In Large Language Model With Model-agnostic Self-decompression

Zhang Zilun, Sun Yutao, Zhao Tiancheng, Sha Leigang, Xu Ruochen, Lee Kyusong, Yin Jianwei. Arxiv 2024

[Paper]    
Fine Tuning Multimodal Models Training Techniques

Humans can retain old knowledge while learning new information, but Large Language Models (LLMs) often suffer from catastrophic forgetting when post-pretrained or supervised fine-tuned (SFT) on domain-specific data. Moreover, for Multimodal Large Language Models (MLLMs) which are composed of the LLM base and visual projector (e.g. LLaVA), a significant decline in performance on language benchmarks was observed compared to their single-modality counterparts. To address these challenges, we introduce a novel model-agnostic self-decompression method, Tree Generation (TG), that decompresses knowledge within LLMs into the training corpus. This paper focuses on TG-SFT, which can synthetically generate SFT data for the instruction tuning steps. By incorporating the dumped corpus during SFT for MLLMs, we significantly reduce the forgetting problem.

Similar Work