Bridging The Gap: Dynamic Learning Strategies For Improving Multilingual Performance In Llms · The Large Language Model Bible Contribute to LLM-Bible

Bridging The Gap: Dynamic Learning Strategies For Improving Multilingual Performance In Llms

Kumar Somnath, Balloli Vaibhav, Ranjit Mercy, Ahuja Kabir, Ganu Tanuja, Sitaram Sunayana, Bali Kalika, Nambi Akshay. Arxiv 2024

[Paper]    
Fine Tuning Pretraining Methods Prompting RAG Training Techniques

Large language models (LLMs) are at the forefront of transforming numerous domains globally. However, their inclusivity and effectiveness remain limited for non-Latin scripts and low-resource languages. This paper tackles the imperative challenge of enhancing the multilingual performance of LLMs without extensive training or fine-tuning. Through systematic investigation and evaluation of diverse languages using popular question-answering (QA) datasets, we present novel techniques that unlock the true potential of LLMs in a polyglot landscape. Our approach encompasses three key strategies that yield significant improvements in multilingual proficiency. First, by meticulously optimizing prompts tailored for polyglot LLMs, we unlock their latent capabilities, resulting in substantial performance boosts across languages. Second, we introduce a new hybrid approach that synergizes LLM Retrieval Augmented Generation (RAG) with multilingual embeddings and achieves improved multilingual task performance. Finally, we introduce a novel learning approach that dynamically selects the optimal prompt strategy, LLM model, and embedding model per query at run-time. This dynamic adaptation maximizes the efficacy of LLMs across languages, outperforming best static and random strategies. Additionally, our approach adapts configurations in both offline and online settings, and can seamlessly adapt to new languages and datasets, leading to substantial advancements in multilingual understanding and generation across diverse languages.

Similar Work