IGOT: Information Gain Optimized Tokenizer On Domain Adaptive Pretraining · The Large Language Model Bible Contribute to LLM-Bible

IGOT: Information Gain Optimized Tokenizer On Domain Adaptive Pretraining

Feng Dawei, Zhang Yihai, Xu Zhixuan. Arxiv 2024

[Paper]    
Fine Tuning GPT Model Architecture Pretraining Methods Training Techniques

Pretrained Large Language Models (LLM) such as ChatGPT, Claude, etc. have demonstrated strong capabilities in various fields of natural language generation. However, there are still many problems when using LLM in specialized domain-specific fields. When using generative AI to process downstream tasks, a common approach is to add new knowledge (e.g., private domain knowledge, cutting-edge information) to a pretrained model through continued training or fine-tuning. However, whether there is a universal paradigm for domain adaptation training is still an open question. In this article, we proposed Information Gain Optimized Tokenizer (IGOT), which analyzes the special token set of downstream tasks, constructs a new subset using heuristic function \(\phi\) with the special token and its information gain, to build new domain-specific tokenizer, and continues pretraining on the downstream task data. We explored the many positive effects of this method’s customized tokenizer on domain-adaptive pretraining and verified this method can perform better than the ordinary method of just collecting data and fine-tuning. Based on our experiment, the continued pretraining process of IGOT with LLaMA-7B achieved 11.9% token saving, 12.2% training time saving, and 5.8% maximum GPU VRAM usage saving, combined with the T5 model, we can even reach a 31.5% of training time saving, making porting general generative AI to specific domains more effective than before. In domain-specific tasks, supervised \(IGOT_\tau\) shows great performance on reducing both the convergence radius and convergence point during keep pretraining.

Similar Work