Generative artificial intelligence (GenAI) is rapidly entering K-12 classrooms worldwide, initiating urgent debates about its potential to either reduce or exacerbate educational inequalities. Drawing on interviews with 30 K-12 teachers across the United States, South Africa, and Taiwan, this study examines how teachers navigate this GenAI tension around educational equalities. We found teachers actively framed GenAI education as an equality-oriented practice: they used it to alleviate pre-existing inequalities while simultaneously working to prevent new inequalities from emerging. Despite these efforts, teachers confronted persistent systemic barriers, i.e., unequal infrastructure, insufficient professional training, and restrictive social norms, that individual initiative alone could not overcome. Teachers thus articulated normative visions for more inclusive GenAI education. By centering teachers’ practices, constraints, and future envisions, this study contributes a global account of how GenAI education is being integrated into K-12 contexts and highlights what is required to make its adoption genuinely equal.
AI-generated non-consensual intimate imagery (AIG-NCII) is an emerging social problem due to the advancement of AI tools. While recent incidents in middle and high schools have highlighted the urgency of this issue, there is limited understanding of what concrete supports schools need to effectively address AIG-NCII. To fill this gap, we conducted an interview study with 20 educators in the U.S. and investigated their attitudes, experiences, and practices related to AIG-NCII. Educators expressed concerns about both students' and their own vulnerability, as AIG-NCII may cause moral decline among students, while educators themselves could become victims. Nevertheless, existing practices in schools are limited, and they lack both training and systematic policies. Challenges such as a lack of resources, unclear legal boundaries, and limited knowledge of AI make implementation difficult. The findings of this paper contribute to interactive educational tool design, curriculum design, and policy-making, especially regarding the need for multi-stakeholder strategies to address issues surrounding AIG-NCII.
While conversations around the future of education focus on AI, robots and VR, they often overlook how children imagine futures within their own educational worlds. We conducted participatory speculative design workshops with 92 students (ages 12–14) in two contrasting settings: an autonomy-oriented international school in Finland and an exam-driven, high-stakes public school in India. Reflexive thematic analysis revealed that pedagogical ecologies, rather than national cultures, shaped children’s technological imaginaries and future orientations. Students in Finland envisioned pragmatic, technologically advanced, yet human centered classrooms, whereas students in India prioritized curricular choice, emotional safety, and systemic fairness. In the workshops in India, we observed that speculative possibilities expanded with careful scaffolding but remained tethered to current realities when scaffolding was disrupted. We argue for "plural design futuring" grounded in children’s lived experiences and contribute methodological insights into scaffolding as a critical condition for participatory future-making. Our findings demonstrate how local educational cultures fundamentally shape the possibilities of speculative design.
Recruiting, retaining, and educating students in computing is a frequent research topic in CHI. However, students' sociotechnical experiences of registering for classes are understudied -- especially those of socioeconomic-diverse students. These experiences matter: research shows that registration problems bring long-term consequences to student successes. We investigate students' socioeconomic status (SES) impact on registration experiences through three studies: a case study with education professionals using an emerging analytic method, SocioeconomicMag (SESMag); interviews with faculty/staff/students from 8 universities; and observations of 14 SES-diverse students registering for classes. Results showed: (1) 5 SES-inclusivity bugs which arose 30 times, 72% more often by lower-SES students than by higher-SES students. (2) 6/7 lower-SES students (but only 2/7 higher-SES students) expected downstream problems from the registration issues. (3) The risk-to-negative-outcomes rate was 3 times higher for lower-SES students.
Generative AI (GenAI) tools are increasingly pervasive, pushing instructors to redesign how students use GenAI tools in coursework. We conceptualize this work as emergency pedagogical design: reactive, indirect efforts by instructors to shape student-AI interactions without control over commercial interfaces. To understand practices of lead users conducting emergency pedagogical design, we conducted interviews (n=13) and a survey (n=169) of computing instructors. These instructors repeatedly encountered five barriers: fragmented buy-in for revising courses; policy crosswinds from non-prescriptive institutional guidance; implementation challenges as instructors attempt interventions; assessment misfit as student-AI interactions are only partially visible to instructors; and lack of resources, including time, staffing, and paid tool access. We use these findings to present emergency pedagogical design as a distinct design setting for HCI and outline recommendations for HCI researchers, academic institutions, and organizations to effectively support instructors in adapting courses to GenAI.
Generative AI is reshaping education, yet most university AI policies are written without students and focus on penalizing misuse. This top-down approach sidelines those most affected from decisions that shape their everyday learning, resulting in confusion and fear about acceptable use. We examine how participatory, student-driven AI policy design can address this disconnect. We report on a three-part workshop series in a graduate design course at a minority-serving university in the U.S., where two student leaders facilitated discussions without faculty present. Eight participants shared candid accounts of their AI use, co-authored ten policy recommendations, and visualized them in a zine that circulated across campus. The resulting policies surfaced concerns absent from top-down governance, such as the double standard of requiring students to disclose or abstain from AI use while faculty face no such expectations. We argue that engaging students in AI governance carries value beyond the resulting policies, and offer transferable strategies for fostering participation across disciplines—a model for calling students in rather than calling students out.
Model documentation plays a crucial role in promoting responsible AI (RAI) development. The paradigm shift from traditional machine learning models to Generative AI (GenAI) models has reshaped the conditions under which documentation is produced, particularly
on open-source platforms where models are hosted and shared. To investigate how this paradigm shift has manifested in developers’ documentation practices, we conducted interviews with 17 GenAI developers who document models on open-source platforms. Our findings illustrated that uncertainties have become the defining feature of developers’ GenAI documentation practices, which unfolds in three interrelated forms: (1) normative and epistemic uncertainties in determining documentation content; (2) methodological
uncertainties in how to evaluate and communicate model properties; and (3) ecosystemic uncertainties in who should document. We argue that the uncertainties in GenAI documentation require coordinated interventions, including infrastructural support to address epistemic and methodological uncertainties, community-based mechanisms to cultivate RAI documentation norms, and collaboration across supply chain actors to address ecosystemic uncertainties.