Live in the Loop: Rapid Run-time Feedback for Prompts

要旨

Programmers formulate prompts for code generation based on their understandings of problem domain and LLM ability to execute instructions. A deficiency in either understanding yields inadequate generated code, requiring programmers to revise their understandings based on deficiencies observed in the generated code. We propose ReFiQ (``Result-first Queries''), a tool that simultaneously offers concrete run-time execution results of multiple generated code variations for one prompt. In an exploratory user study (n=8), we observed that programmers systematically compared those results to identify issues in their understandings, deliberately formulated single prompts to explore a range of options of a domain concept, and were more often encouraged to make a deliberate decision informed by their observations instead of directly applying generated code. Participants in our study required fewer iterations to arrive at a satisfying solution in ReFiQ than in our baseline, GitHub Copilot.

著者
Toni Mattis
University of Potsdam, Potsdam, Germany
Abdullatif Ghajar
University of Potsdam, Potsdam, Germany
Tom Beckmann
University of Potsdam, Potsdam, Germany
Robert Hirschfeld
University of Potsdam, Potsdam, Germany

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Interactive Prompting, Chaining, and LLM Orchestration Tools

P1 - Room 132
7 件の発表
2026-04-13 20:15:00
2026-04-13 21:45:00