Tell Your Model Where To Attend: Post-hoc Attention Steering For Llms · The Large Language Model Bible Contribute to LLM-Bible

Tell Your Model Where To Attend: Post-hoc Attention Steering For Llms

Zhang Qingru, Singh Chandan, Liu Liyuan, Liu Xiaodong, Yu Bin, Gao Jianfeng, Zhao Tuo. Arxiv 2023

[Paper] [Code]    
Attention Mechanism Has Code Model Architecture Prompting RAG Reinforcement Learning

In human-written articles, we often leverage the subtleties of text style, such as bold and italics, to guide the attention of readers. These textual emphases are vital for the readers to grasp the conveyed information. When interacting with large language models (LLMs), we have a similar need - steering the model to pay closer attention to user-specified information, e.g., an instruction. Existing methods, however, are constrained to process plain text and do not support such a mechanism. This motivates us to introduce PASTA

Similar Work