Clues Before Answers: Generation-enhanced Multiple-choice QA · The Large Language Model Bible Contribute to LLM-Bible

Clues Before Answers: Generation-enhanced Multiple-choice QA

Huang Zixian, Wu Ao, Zhou Jiaying, Gu Yu, Zhao Yue, Cheng Gong. Arxiv 2022

[Paper]    
Applications RAG Reinforcement Learning Tools

A trending paradigm for multiple-choice question answering (MCQA) is using a text-to-text framework. By unifying data in different tasks into a single text-to-text format, it trains a generative encoder-decoder model which is both powerful and universal. However, a side effect of twisting a generation target to fit the classification nature of MCQA is the under-utilization of the decoder and the knowledge that can be decoded. To exploit the generation capability and underlying knowledge of a pre-trained encoder-decoder model, in this paper, we propose a generation-enhanced MCQA model named GenMC. It generates a clue from the question and then leverages the clue to enhance a reader for MCQA. It outperforms text-to-text models on multiple MCQA datasets.

Similar Work