Cached Transformers: Improving Transformers With Differentiable Memory Cache · The Large Language Model Bible Contribute to LLM-Bible

Cached Transformers: Improving Transformers With Differentiable Memory Cache

Zhang Zhaoyang, Shao Wenqi, Ge Yixiao, Wang Xiaogang, Gu Jinwei, Luo Ping. Arxiv 2023

[Paper]    
Applications Attention Mechanism Language Modeling Model Architecture Pretraining Methods Training Techniques Transformer

This work introduces a new Transformer model called Cached Transformer, which uses Gated Recurrent Cached (GRC) attention to extend the self-attention mechanism with a differentiable memory cache of tokens. GRC attention enables attending to both past and current tokens, increasing the receptive field of attention and allowing for exploring long-range dependencies. By utilizing a recurrent gating unit to continuously update the cache, our model achieves significant advancements in \textbf{six} language and vision tasks, including language modeling, machine translation, ListOPs, image classification, object detection, and instance segmentation. Furthermore, our approach surpasses previous memory-based techniques in tasks such as language modeling and displays the ability to be applied to a broader range of situations.

Similar Work