Unsupervised Evaluation Of Code Llms With Round-trip Correctness · The Large Language Model Bible Contribute to LLM-Bible

Unsupervised Evaluation Of Code Llms With Round-trip Correctness

Allamanis Miltiadis, Panthaplackel Sheena, Yin Pengcheng. Arxiv 2024

[Paper]    
Reinforcement Learning

To evaluate code large language models (LLMs), research has relied on a few small manually curated benchmarks, such as HumanEval and MBPP, which represent a narrow part of the real-world software domains. In this work, we introduce round-trip correctness (RTC) as an alternative evaluation method. RTC allows Code LLM evaluation on a broader spectrum of real-world software domains without the need for costly human curation. RTC rests on the idea that we can ask a model to make a prediction (e.g., describe some code using natural language), feed that prediction back (e.g., synthesize code from the predicted description), and check if this round-trip leads to code that is semantically equivalent to the original input. We show how to employ RTC to evaluate code synthesis and editing. We find that RTC strongly correlates with model performance on existing narrow-domain code synthesis benchmarks while allowing us to expand to a much broader set of domains and tasks which was not previously possible without costly human annotations.

Similar Work