Surgment: Segmentation-enabled Semantic Search and Creation of Visual Question and Feedback to Support Video-Based Surgery Learning

要旨

Videos are prominent learning materials to prepare surgical trainees before they enter the operating room (OR). In this work, we explore techniques to enrich the video-based surgery learning experience. We propose Surgment, a system that helps expert surgeons create exercises with feedback based on surgery recordings. Surgment is powered by a few-shot-learning-based pipeline (SegGPT+SAM) to segment surgery scenes, achieving an accuracy of 92\%. The segmentation pipeline enables functionalities to create visual questions and feedback desired by surgeons from a formative study. Surgment enables surgeons to 1) retrieve frames of interest through sketches, and 2) design exercises that target specific anatomical components and offer visual feedback. In an evaluation study with 11 surgeons, participants applauded the search-by-sketch approach for identifying frames of interest and found the resulting image-based questions and feedback to be of high educational value.

著者
Jingying Wang
University of Michigan, Ann Arbor, Michigan, United States
Haoran Tang
University of Michigan, Ann Arbor, Michigan, United States
Taylor Kantor
University of Michigan, Ann Arbor, Michigan, United States
Tandis Soltani
University of Michigan, Ann Arbor, Michigan, United States
Vitaliy Popov
University of Michigan, Ann Arbor, Michigan, United States
Xu Wang
University of Michigan, Ann Arbor, Michigan, United States
論文URL

doi.org/10.1145/3613904.3642587

動画

会議: CHI 2024

The ACM CHI Conference on Human Factors in Computing Systems (https://chi2024.acm.org/)

セッション: Healthcare Training

318B
5 件の発表
2024-05-14 20:00:00
2024-05-14 21:20:00