RAG Without the Lag: Enabling "What-If" Analysis for Retrieval-Augmented Generation Pipelines

要旨

Retrieval-augmented generation (RAG) pipelines have become the de-facto approach for building AI assistants with external knowledge. Given a user query, RAG pipelines retrieve (R) information from external sources, before invoking a Large Language Model (LLM), augmented (A) with this information, to generate (G) responses. However, developing effective RAG pipelines is challenging because retrieval and generation components---often chained in varying orders---are intertwined, making it hard to identify which component(s) cause errors in the output. Developers often need to answer "what-if" questions---e.g., what if chunk sizes were larger or retrieval used embeddings versus keywords---but such experimentation requires hours of re-processing. We present RAGGY, a developer tool that enables rapid "what-if" analysis by combining a Python library of composable RAG primitives with an interactive debugging interface. We contribute the design and implementation of RAGGY, insights into expert debugging patterns through a qualitative study with 12 engineers, and design implications for RAG tools.

受賞
Best Paper
著者
Quentin Romero Lauro
University of Pittsburgh, Pittsburgh, Pennsylvania, United States
Shreya Shankar
University of California, Berkeley, Berkeley, California, United States
Sepanta Zeighami
University of California Berkeley, Berkeley, California, United States
Aditya Parameswaran
UC Berkeley, Berkeley, California, United States

会議: CHI 2026

ACM CHI Conference on Human Factors in Computing Systems

セッション: Conversational AI

P1 - Room 125
7 件の発表
2026-04-14 18:00:00
2026-04-14 19:30:00