Cracking the Code: Evaluating Zero-Shot Prompting Methods for Providing Programming Feedback

Published: 06 Mar 2025, Last Modified: 27 Mar 2025ICLR-25 HAIC WorkshopEveryoneRevisionsBibTeXCC BY 4.0
Track: tiny / short paper (up to 5 pages)
Keywords: Prompt Engineering, Feedback Evaluation, R Programming Education
TL;DR: We developed a systematic evaluation framework for assessing LLM responses on programming errors and conducted a case study.
Abstract: We introduce an evaluation framework to assess the feedback given by large language models (LLMs) under different prompt engineering techniques and conduct a case study, systematically varying prompts to examine their influence on feedback quality for common programming errors in R. Our findings suggest that prompts recommending a stepwise approach improve precision, whereas omitting explicit details on which data to analyze can bolster error identification.
Submission Number: 12
Loading

OpenReview is a long-term project to advance science through improved peer review with legal nonprofit status. We gratefully acknowledge the support of the OpenReview Sponsors. © 2025 OpenReview