Validating AI-Generated Code with Live Programming

要旨

AI-powered programming assistants are increasingly gaining popularity, with GitHub Copilot alone used by over a million developers worldwide. These tools are far from perfect, however, producing code suggestions that may be incorrect in subtle ways. As a result, developers face a new challenge: validating AI's suggestions. This paper explores whether Live Programming (LP), a continuous display of a program's runtime values, can help address this challenge. To answer this question, we built a Python editor that combines an AI-powered programming assistant with an existing LP environment. Using this environment in a between-subjects study (N=17), we found that by lowering the cost of validation by execution, LP can mitigate over- and under-reliance on AI-generated programs and reduce the cognitive load of validation for certain types of tasks.

著者
Kasra Ferdowsi
UC San Diego, La Jolla, California, United States
Ruanqianqian (Lisa) Huang
UC San Diego, La Jolla, California, United States
Michael B.. James
University of California, San Diego, San Diego, California, United States
Nadia Polikarpova
University of California, San Diego, La Jolla, California, United States
Sorin Lerner
UC San Diego, La Jolla, California, United States
論文URL

doi.org/10.1145/3613904.3642495

動画

会議: CHI 2024

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2024.acm.org/)

セッション: Coding with AI

324
5 件の発表
2024-05-14 23:00:00
2024-05-15 00:20:00