Llava-read: Enhancing Reading Ability Of Multimodal Language Models · The Large Language Model Bible Contribute to LLM-Bible

Llava-read: Enhancing Reading Ability Of Multimodal Language Models

Zhang Ruiyi, Zhou Yufan, Chen Jian, Gu Jiuxiang, Chen Changyou, Sun Tong. Arxiv 2024

[Paper]    
Fine Tuning Multimodal Models

Large multimodal language models have demonstrated impressive capabilities in understanding and manipulating images. However, many of these models struggle with comprehending intensive textual contents embedded within the images, primarily due to the limited text recognition and layout understanding ability. To understand the sources of these limitations, we perform an exploratory analysis showing the drawbacks of classical visual encoders on visual text understanding. Hence, we present LLaVA-Read, a multimodal large language model that utilizes dual visual encoders along with a visual text encoder. Our model surpasses existing state-of-the-art models in various text-rich image understanding tasks, showcasing enhanced comprehension of textual content within images. Together, our research suggests visual text understanding remains an open challenge and an efficient visual text encoder is crucial for future successful multimodal systems.

Similar Work