Exploring Multimodal Generative AI for Education through Co-design Workshops with Students

要旨

Multimodal large language models (MLLMs) are Generative AI models that take different modalities such as text, audio, and video as input and generate appropriate multimodal output. Since such models will be integrated into future educational tools, a human-centered design approach that takes students’ perspectives into account is essential while designing such applications. This paper describes two co-design workshops which were conducted with 79 student groups to examine how they design and prototype future educational tools integrated with MLLMs. Through various activities in the workshops, students discussed relevant educational problems, created journey maps, storyboards and low fidelity prototypes for their applications, and evaluated their applications based on relevant design principles. We found that students’ applications used MLLMs for important learning environment design features such as multimodal content creation, personalization, and feedback. Based on these findings, we discuss future research directions for the design of multimodality in generative AI educational applications.

著者
Prajish Prasad
FLAME University, Pune, India
Rishabh Balse
University of Cambridge, Cambridge, United Kingdom
Dhwani Balchandani
FLAME University, Pune, India
DOI

10.1145/3706598.3714146

論文URL

https://dl.acm.org/doi/10.1145/3706598.3714146

動画

会議: CHI 2025

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2025.acm.org/)

セッション: Classroom Technology

G302
7 件の発表
2025-04-29 23:10:00
2025-04-30 00:40:00
日本語まとめ
読み込み中…