Scratch Copilot Evaluation: Assessing Ai-assisted Creative Coding For Families · The Large Language Model Bible Contribute to LLM-Bible

Scratch Copilot Evaluation: Assessing Ai-assisted Creative Coding For Families

Druga Stefania, Otero Nancy. Arxiv 2023

[Paper]    
Applications Reinforcement Learning Tools

How can AI enhance creative coding experiences for families? This study explores the potential of large language models (LLMs) in helping families with creative coding using Scratch. Based on our previous user study involving a prototype AI assistant, we devised three evaluation scenarios to determine if LLMs could help families comprehend game code, debug programs, and generate new ideas for future projects. We utilized 22 Scratch projects for each scenario and generated responses from LLMs with and without practice tasks, resulting in 120 creative coding support scenario datasets. In addition, the authors independently evaluated their precision, pedagogical value, and age-appropriate language. Our findings show that LLMs achieved an overall success rate of more than 80% on the different tasks and evaluation criteria. This research offers valuable information on using LLMs for creative family coding and presents design guidelines for future AI-supported coding applications. Our evaluation framework, together with our labeled evaluation data, is publicly available.

Similar Work