Crowds and Collaboration

会議の名前
CSCW2021
Latexify Math: Mathematical Formula Markup Revision to Assist Collaborative Editing in Math Q&A Sites
要旨

Collaborative editing questions and answers plays an important role in quality control of Mathematics Stack Exchange which is a math Q&A Site. Our study of post edits in Mathematics Stack Exchange shows that there is a large number of math-related edits about latexifying formulas, revising LaTeX and converting the blurred math formula screenshots to LaTeX sequence. Despite its importance, manually editing one math-related post especially those with complex mathematical formulas is time-consuming and error-prone even for experienced users. To assist post owners and editors to do this editing, we have developed an edit-assistance tool, MathLatexEdit for formula latexification, LaTeX revision and screenshot transcription. We formulate this formula editing task as a translation problem, in which an original post is translated to a revised post. MathLatexEdit implements a deep learning based approach including two encoder-decoder models for textual and visual LaTeX edit recommendation with math-specific inference. The two models are trained on large-scale historical original-edited post pairs and synthesized screenshot-formula pairs. Our evaluation of MathLatexEdit not only demonstrates the accuracy of our model, but also the usefulness of MathLatexEdit in editing real-world posts which are accepted in Mathematics Stack Exchange.

著者
Suyu Ma
Monash University, Melbourne, VIC, Australia
Chunyang Chen
Monash University, Melbourne, Victoria, Australia
Hourieh Khalajzadeh
Monash University, Melbourne, VIC, Australia
John Grundy
Monash University, Melbourne, Victoria, Australia
論文URL

https://doi.org/10.1145/3479547

動画
A Multi-platform Study of Crowd Signals Associated with Successful Online Fundraising
要旨

The growing popularity of online fundraising (aka “crowdfunding”) has attracted significant research on the subject. In contrast to previous studies that attempt to predict the success of crowdfunded projects based on specific characteristics of the projects and their creators, we present a more general approach that focuses on crowd dynamics and is robust to the particularities of different crowdfunding platforms. We rely on a multi-level analysis to investigate the correlates, predictive importance, and quasi-causal effects of features that describe crowd dynamics in determining the success of crowdfunded projects. By applying a multi-level analysis to a study of fundraising in three different online markets, we uncover general crowd dynamics that ultimately decide which projects will succeed. In all levels of analysis and across the three different platforms, we consistently find that funders’ behavioural signals (1) are significantly correlated with fundraising success;(2) approximate fundraising outcomes better than the characteristics of projects and their creators such as credit grade, company valuation, and subject domain; and (3) have significant quasi-causal effects on fundraising outcomes while controlling for potentially confounding project variables. By showing that universal features deduced from crowd behaviour are predictive of fundraising success on different crowdfunding platforms, our work provides design-relevant insights about novel types of collective decision-making online. This research inspires thus potential ways to leverage cues from the crowd and catalyses research into crowd-aware system design.

著者
Henry Kudzanai. Dambanemuya
Northwestern University, Evanston, Illinois, United States
Emoke Agnes Horvat
Northwestern University, Evanston, Illinois, United States
論文URL

https://doi.org/10.1145/3449189

動画
Discovering and Validating AI Errors With Crowdsourced Failure Reports
要旨

AI systems can fail to learn important behaviors, leading to real-world issues like safety concerns and biases. Discovering these systematic failures often requires significant developer attention, from hypothesizing potential edge cases to collecting evidence and validating patterns. To scale and streamline this process, we introduce crowdsourced failure reports, end-user descriptions of how or why a model failed, and show how developers can use them to detect AI errors. We also design and implement Deblinder, a visual analytics system for synthesizing failure reports that developers can use to discover and validate systematic failures. In semi-structured interviews and think-aloud studies with 10 AI practitioners, we explore the affordances of the Deblinder system and the applicability of failure reports in real-world settings. Lastly, we show how collecting additional data from the groups identified by developers can improve model performance.

著者
Ángel Alexander Cabrera
Carnegie Mellon University, Pittsburgh, Pennsylvania, United States
Abraham J. Druck
Carnegie Mellon University, Pittsburgh, Pennsylvania, United States
Jason I. Hong
Carnegie Mellon University, Pittsburgh, Pennsylvania, United States
Adam Perer
Carnegie Mellon University, Pittsburgh, Pennsylvania, United States
論文URL

https://doi.org/10.1145/3479569

動画
CrowdFolio: Understanding How Holistic and Decomposed Workflows Influence Feedback on Online Portfolios
要旨

Freelancers increasingly earn their livelihood through online marketplaces. To attract new clients, freelancers continuously curate their online portfolios to convey their unique skills and style. However, many lack access to rapid, regular, and inexpensive feedback needed to improve their portfolios. Existing crowd feedback systems, which collect feedback on individual creative projects (i.e., decomposed approach), could fill this need, but it is unclear how they might support feedback on multiple projects (i.e., holistic approach). In a between-subjects study with 30 freelancers, we compared decomposed and holistic feedback collection approaches using CrowdFolio, a crowd feedback system for portfolios. The holistic approach helped freelancers discover new ways to describe their work, while the decomposed approach provided detailed insight about the visual attractiveness of projects. This study contributes evidence that portfolio feedback systems, regardless of collection approach, can positively support professional development by impacting how freelancers portray themselves online and reflect on their identity.

著者
Eureka Foong
University of Tokyo, Tokyo, Japan
Joy O. Kim
Adobe Research, San Francisco, California, United States
Mira Dontcheva
Adobe Research, Seattle, Washington, United States
Elizabeth Gerber
Northwestern University, Evanston, Illinois, United States
論文URL

https://doi.org/10.1145/3449096

動画
The Challenge of Variable Effort Crowdsourcing & How Visible Gold Can Help
要旨

We consider a class of variable effort human annotation tasks in which the number of labels required per item can greatly vary (e.g., finding all faces in an image, named entities in a text, bird calls in an audio recording, etc.). In such tasks, some items require far more effort than others to annotate. Furthermore, the per-item annotation effort is not known until after each item is annotated since determining the number of labels required is an implicit part of the annotation task itself. On an image bounding-box task with crowdsourced annotators, we show that annotator accuracy and recall consistently drop as effort increases. We hypothesize reasons for this drop and investigate a set of approaches to counteract it. Firstly, we benchmark on this task a set of general best-practice methods for quality crowdsourcing. Notably, only one of these methods actually improves quality: the use of visible gold questions that provide periodic feedback to workers on their accuracy as they work. Given these promising results, we then investigate and evaluate variants of the visible gold approach, yielding further improvement. Final results show a 7% improvement in bounding-box accuracy over the baseline. We discuss the generality of the visible gold approach and promising directions for future research.

著者
Danula Hettiachchi
Amazon, Seattle, Washington, United States
Mike Schaekermann
Amazon, Toronto, Ontario, Canada
Tristan J. McKinney
Amazon, Palo Alto, California, United States
Matthew Lease
Amazon, Seattle, Washington, United States
論文URL

https://doi.org/10.1145/3476073

動画
Goldilocks: Consistent Crowdsourced Scalar Annotations with Relative Uncertainty
要旨

Human ratings have become a crucial resource for training and evaluating machine learning systems. However, traditional elicitation methods for absolute and comparative rating suffer from issues with consistency and often do not distinguish between uncertainty due to disagreement between annotators and ambiguity inherent to the item being rated. In this work, we present Goldilocks, a novel crowd rating elicitation technique for collecting calibrated scalar annotations that also distinguishes inherent ambiguity from inter-annotator disagreement. We introduce two main ideas: grounding absolute rating scales with examples and using a two-step bounding process to establish a range for an item's placement}. We test our designs in three domains: judging toxicity of online comments, estimating satiety of food depicted in images, and estimating age based on portraits. We show that (1) Goldilocks can improve consistency in domains where interpretation of the scale is not universal, and that (2) representing items with ranges lets us simultaneously capture different sources of uncertainty leading to better estimates of pairwise relationship distributions.

著者
Quan Ze Chen
University of Washington, Seattle, Washington, United States
Daniel S. Weld
University of Washington, Seattle, Washington, United States
Amy X.. Zhang
University of Washington, Seattle, Washington, United States
論文URL

https://doi.org/10.1145/3476076

Task Execution Quality Maximization for Mobile Crowdsourcing in Geo-Social Networks
要旨

With the rapid development of smart devices and high-quality wireless technologies, mobile crowdsourcing (MCS) has been drawing increasing attention with its great potential in collaboratively completing complicated tasks on a large scale. A key issue toward successful MCS is participant recruitment, where a MCS platform directly recruits suitable crowd participants to execute outsourced tasks by physically traveling to specified locations. Recently, a novel recruitment strategy, namely Word-of-Mouth(WoM)-based MCS, has emerged to effectively improve recruitment effectiveness, by fully exploring users' mobility traces and social relationships on geo-social networks. Against this background, we study in this paper a novel problem, namely Expected Task Execution Quality Maximization (ETEQM) for MCS in geo-social networks, which strives to search a subset of seed users to maximize the expected task execution quality of all recruited participants, under a given incentive budget. To characterize the MCS task propagation process over geo-social networks, we first adopt a propagation tree structure to model the autonomous recruitment between the referrers and the referrals. Based on the model, we then formalize the task execution quality and devise a novel incentive mechanism by harnessing the business strategy of multi-level marketing. We formulate our ETEQM problem as a combinatorial optimization problem, and analyze its NP hardness and high-dimensional characteristics. Based on a cooperative co-evolution framework, we proposed a divide-and-conquer problem-solving approach named ETEQM-CC. We conduct extensive simulation experiments and a case study, verifying the effectiveness of our proposed approach.

著者
Liang Wang
Northwestern Polytechnical University, Xi'an, China
Zhiwen Yu
Northwestern Polytechnical University, Xi'an, China
Dingqi Yang
University of Fribourg, Fribourg, Switzerland
TIAN WANG
Huaqiao University, Xiamen, Fujian, China
En Wang
Jilin University, Changchun, China
Bin Guo
northwestern polytechnical univ., xian, China
Daqing Zhang
Peking University, Beijing, China
論文URL

https://doi.org/10.1145/3476053

動画
Task Assignment Strategies for Crowd Worker Ability Improvement
要旨

Workers are the most important resource in crowdsourcing. However, only investing in worker-centric needs, such as skill improvement, often conflicts with short-term platform-centric needs, such as task throughput. This paper studies learning strategies in task assignment in crowdsourcing and their impact on platform-centric needs. We formalize learning potential of individual tasks and collaborative tasks, and devise an iterative task assignment and completion approach that implements strategies grounded in learning theories. We conduct experiments to compare several learning strategies in terms of skill improvement, and in terms of task throughput and contribution quality. We discuss how our findings open new research directions in learning and collaboration.

著者
Masaki Matsubara
University of Tsukuba, Tsukuba, Japan
Ria M. Borromeo
University of the Philippines, Diliman, Philippines
Amer-Yahia Sihem
CNRS/UGA, Grenoble, France
Atsuyuki Morishima
University of Tsukuba, Tsukuba, Japan
論文URL

https://doi.org/10.1145/3479519

動画