Fairytalecqa: Integrating A Commonsense Knowledge Graph Into Children's Storybook Narratives · The Large Language Model Bible Contribute to LLM-Bible

Fairytalecqa: Integrating A Commonsense Knowledge Graph Into Children's Storybook Narratives

Chen Jiaju, Lu Yuxuan, Zhang Shao, Yao Bingsheng, Dong Yuanzhe, Xu Ying, Li Yunyao, Wang Qianwen, Wang Dakuo, Sun Yuling. Arxiv 2023

[Paper]    
Applications GPT Model Architecture Prompting Reinforcement Learning

AI models (including LLM) often rely on narrative question-answering (QA) datasets to provide customized QA functionalities to support downstream children education applications; however, existing datasets only include QA pairs that are grounded within the given storybook content, but children can learn more when teachers refer the storybook content to real-world knowledge (e.g., commonsense knowledge). We introduce the FairytaleCQA dataset, which is annotated by children education experts, to supplement 278 storybook narratives with educationally appropriate commonsense knowledge. The dataset has 5,868 QA pairs that not only originate from the storybook narrative but also contain the commonsense knowledge grounded by an external knowledge graph (i.e., ConceptNet). A follow-up experiment shows that a smaller model (T5-large) fine-tuned with FairytaleCQA reliably outperforms much larger prompt-engineered LLM (e.g., GPT-4) in this new QA-pair generation task (QAG). This result suggests that: 1) our dataset brings novel challenges to existing LLMs, and 2) human experts’ data annotation are still critical as they have much nuanced knowledge that LLMs do not know in the children educational domain.

Similar Work