Guided Transformer: Leveraging Multiple External Sources For Representation Learning In Conversational Search · The Large Language Model Bible Contribute to LLM-Bible

Guided Transformer: Leveraging Multiple External Sources For Representation Learning In Conversational Search

Hashemi Helia, Zamani Hamed, Croft W. Bruce. Arxiv 2020

[Paper]    
Applications Attention Mechanism Model Architecture Pretraining Methods RAG Transformer

Asking clarifying questions in response to ambiguous or faceted queries has been recognized as a useful technique for various information retrieval systems, especially conversational search systems with limited bandwidth interfaces. Analyzing and generating clarifying questions have been studied recently but the accurate utilization of user responses to clarifying questions has been relatively less explored. In this paper, we enrich the representations learned by Transformer networks using a novel attention mechanism from external information sources that weights each term in the conversation. We evaluate this Guided Transformer model in a conversational search scenario that includes clarifying questions. In our experiments, we use two separate external sources, including the top retrieved documents and a set of different possible clarifying questions for the query. We implement the proposed representation learning model for two downstream tasks in conversational search; document retrieval and next clarifying question selection. Our experiments use a public dataset for search clarification and demonstrate significant improvements compared to competitive baselines.

Similar Work