Llm-oriented Retrieval Tuner · The Large Language Model Bible Contribute to LLM-Bible

Llm-oriented Retrieval Tuner

Sun Si, Zhang Hanqing, Liu Zhiyuan, Bao Jie, Song Dawei. Arxiv 2024

[Paper]    
Applications GPT Language Modeling Model Architecture

Dense Retrieval (DR) is now considered as a promising tool to enhance the memorization capacity of Large Language Models (LLM) such as GPT3 and GPT-4 by incorporating external memories. However, due to the paradigm discrepancy between text generation of LLM and DR, it is still an open challenge to integrate the retrieval and generation tasks in a shared LLM. In this paper, we propose an efficient LLM-Oriented Retrieval Tuner, namely LMORT, which decouples DR capacity from base LLM and non-invasively coordinates the optimally aligned and uniform layers of the LLM towards a unified DR space, achieving an efficient and effective DR without tuning the LLM itself. The extensive experiments on six BEIR datasets show that our approach could achieve competitive zero-shot retrieval performance compared to a range of strong DR models while maintaining the generation ability of LLM.

Similar Work