Never Miss A Beat: An Efficient Recipe For Context Window Extension Of Large Language Models With Consistent "middle" Enhancement · The Large Language Model Bible Contribute to LLM-Bible

Never Miss A Beat: An Efficient Recipe For Context Window Extension Of Large Language Models With Consistent "middle" Enhancement

Wu Tong, Zhao Yanpeng, Zheng Zilong. Arxiv 2024

[Paper]    
Fine Tuning Pretraining Methods RAG Training Techniques

Recently, many methods have been developed to extend the context length of pre-trained large language models (LLMs), but they often require fine-tuning at the target length (\(\gg4K\)) and struggle to effectively utilize information from the middle part of the context. To address these issues, we propose \(\textbf{C}\)ontinuity-\(\textbf{R}\)elativity ind\(\textbf{E}\)xing with g\(\textbf{A}\)ussian \(\textbf{M}\)iddle (CREAM), which interpolates positional encodings by manipulating position indices. Apart from being simple, CREAM is training-efficient: it only requires fine-tuning at the pre-trained context window (eg, Llama 2-4K) and can extend LLMs to a much longer target context length (eg, 256K). To ensure that the model focuses more on the information in the middle, we introduce a truncated Gaussian to encourage sampling from the middle part of the context during fine-tuning, thus alleviating the Lost-in-the-Middle'' problem faced by long-context LLMs. Experimental results show that CREAM successfully extends LLMs to the target length for both Base and Chat versions of \\(\texttt\{Llama2-7B\}\\) withNever Miss A Beat’’. Our code will be publicly available soon.

Similar Work