skip to main content
10.1145/3078714.3078715acmconferencesArticle/Chapter ViewAbstractPublication PageshtConference Proceedingsconference-collections
research-article

Clarity is a Worthwhile Quality: On the Role of Task Clarity in Microtask Crowdsourcing

Published:04 July 2017Publication History

ABSTRACT

Workers of microtask crowdsourcing marketplaces strive to find a balance between the need for monetary income and the need for high reputation. Such balance is often threatened by poorly formulated tasks, as workers attempt their execution despite a sub-optimal understanding of the work to be done.

In this paper we highlight the role of clarity as a characterising property of tasks in crowdsourcing. We surveyed 100 workers of the CrowdFlower platform to verify the presence of issues with task clarity in crowdsourcing marketplaces, reveal how crowd workers deal with such issues, and motivate the need for mechanisms that can predict and measure task clarity. Next, we propose a novel model for task clarity based on the goal and role clarity constructs. We sampled 7.1K tasks from the Amazon mTurk marketplace, and acquired labels for task clarity from crowd workers. We show that task clarity is coherently perceived by crowd workers, and is affected by the type of the task. We then propose a set of features to capture task clarity, and use the acquired labels to train and validate a supervised machine learning model for task clarity prediction. Finally, we perform a long-term analysis of the evolution of task clarity on Amazon mTurk, and show that clarity is not a property suitable for temporal characterisation.

References

  1. Omar Alonso and Ricardo Baeza-Yates. 2011. Design and implementation of relevance assessments using crowdsourcing. In ECIR. Springer, 153--164. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Omar Alonso, Catherine Marshall, and Marc Najork. 2014. Crowdsourcing a subjective labeling task: a human-centered framework to ensure reliable results. Technical Report. MSR-TR-2014-91.Google ScholarGoogle Scholar
  3. Janine Berg. 2016. Income security in the on-demand economy: findings and policy lessons from a survey of crowdworkers. Comparative Labor Law & Policy Journal 37, 3 (2016).Google ScholarGoogle Scholar
  4. Hein Broekkamp, Bernadette HAM van Hout-Wolters, Gert Rijlaarsdam, and Huub van den Bergh. 2002. Importance in instructional text: teachers' and students' perceptions of task demands. Journal of Educational Psychology 94, 2 (2002), 260.Google ScholarGoogle ScholarCross RefCross Ref
  5. Francisco Cano and María Cardelle-Elawar. 2004. An integrated analysis of secondary school studentsfi conceptions and beliefs about learning. European Journal of Psychology of Education 19, 2 (2004), 167--187.Google ScholarGoogle ScholarCross RefCross Ref
  6. Jeanne Sternlicht Chall and Edgar Dale. 1995. Readability revisited: The new Dale-Chall readability formula. Brookline Books.Google ScholarGoogle Scholar
  7. Kevyn Collins-Thompson. 2014. Computational assessment of text readability: A survey of current and future research. ITL-International Journal of Applied Linguistics 165, 2 (2014), 97--135.Google ScholarGoogle ScholarCross RefCross Ref
  8. Kevyn Collins-Thompson and James P Callan. 2004. A language modeling ap- proach to predicting reading difficulty.. In HLT-NAACL . 193--200.Google ScholarGoogle Scholar
  9. Scott A Crossley, Kristopher Kyle, and Danielle S McNamara. 2015. The tool for the automatic analysis of text cohesion (TAACO): automatic assessment of local, global, and text cohesion. Behavior research methods (2015), 1--11.Google ScholarGoogle Scholar
  10. Tove I Dahl, Margrethe Bals, and Anne Lene Turi. 2005. Are students' beliefs about knowledge and learning associated with their reported use of learning strategies? British journal of educational psychology 75, 2 (2005), 257--273.Google ScholarGoogle Scholar
  11. Edgar Dale and Jeanne S Chall. 1949. The concept of readability. Elementary English 26, 1 (1949), 19--26.Google ScholarGoogle Scholar
  12. Orphée De Clercq, Véronique Hoste, Bart Desmet, Philip Van Oosten, Martine De Cock, and Lieve Macken. 2014. Using the crowd for readability prediction. Natural Language Engineering 20, 03 (2014).Google ScholarGoogle Scholar
  13. Djellel Eddine Difallah, Michele Catasta, Gianluca Demartini, Panagiotis G Ipeiro- tis, and Philippe Cudré-Mauroux. 2015. The Dynamics of Micro-Task Crowdsourcing: The Case of Amazon MTurk. In WWW. International World Wide Web Conferences Steering Committee, 238--247. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Ujwal Gadiraju, Ricardo Kawase, and Stefan Dietze. 2014. A taxonomy of micro- tasks on the web. In Hypertext. ACM, 218--223. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Ujwal Gadiraju, Ricardo Kawase, Stefan Dietze, and Gianluca Demartini. 2015. Understanding malicious behavior in crowdsourcing platforms: the case of online surveys. In CHI. ACM, 1631--1640. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Catherine Grady and Matthew Lease. 2010. Crowdsourcing document relevance assessment with mechanical turk. In HLT-NAACL workshop on creating speech and language data with Amazon's mechanical turk. Association for Computation- alLinguistics, 172--179. Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. Arthur C Graesser, Danielle S McNamara, Max M Louwerse, and Zhiqiang Cai. 2004. Coh-Metrix: analysis of text on cohesion and language. Behavior research methods, instruments, & computers 36, 2 (2004), 193--202.Google ScholarGoogle Scholar
  18. Allison Hadwin. 2006. Student task understanding. In Learning and Teaching Conference. University of Victoria, Victoria, British Columbia, Canada.Google ScholarGoogle Scholar
  19. AF Hadwin, M Oshige, M Miller, and P Wild. 2009. Examining student and instructor task perceptions in a complex engineering design task. In international conference on innovation and practices in engineering design and engineering education. McMaster University, Hamilton, ON, Canada.Google ScholarGoogle Scholar
  20. T Hoßfeld, Raimund Schatz, and Sebastian Egger. 2011. SOS: The MOS is not enough!. In QoMEX. IEEE, 131--136.Google ScholarGoogle Scholar
  21. Lilly C Irani and M Silberman. 2013. Turkopticon: Interrupting worker invisibility in amazon mechanical turk. In CHI. ACM, 611--620. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Diane Lee Jamieson-Noel. 2004. Exploring task definition as a facet of self-regulated learning . Ph.D. Dissertation. Faculty of Education-Simon Fraser University.Google ScholarGoogle Scholar
  23. Rohit J Kate, Xiaoqiang Luo, Siddharth Patwardhan, Martin Franz, Radu Flo- rian, Raymond J Mooney, Salim Roukos, and Chris Welty. 2010. Learning to predict readability using diverse linguistic features. In ACL. Association for Computational Linguistics, 546--554. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Shashank Khanna, Aishwarya Ratan, James Davis, and William Thies. 2010. Evaluating and improving the usability of Mechanical Turk for low-income workers in India. In DEV. ACM, 12. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. J Peter Kincaid, Robert P Fishburne Jr, Richard L Rogers, and Brad S Chissom. 1975. Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel. Technical Report. DTIC Document.Google ScholarGoogle Scholar
  26. Aniket Kittur, Ed H Chi, and Bongwon Suh. 2008. Crowdsourcing user studies with Mechanical Turk. In CHI. ACM, 453--456. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Aniket Kittur, Jeffrey V Nickerson, Michael Bernstein, Elizabeth Gerber, Aaron Shaw, John Zimmerman, Matt Lease, and John Horton. 2013. The future of crowd work. In CSCW. ACM, 1301--1318. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Lieve Luyten, Joost Lowyck, and Francis Tuerlinckx. 2001. Task perception as a mediating variable: A contribution to the validation of instructional knowledge. British Journal of Educational Psychology 71, 2 (2001), 203--223.Google ScholarGoogle ScholarCross RefCross Ref
  29. David Malvern and Brian Richards. 2012. Measures of lexical richness. The Encyclopedia of Applied Linguistics (2012).Google ScholarGoogle Scholar
  30. Catherine C Marshall and Frank M Shipman. 2013. Experiences surveying the crowd: Reflections on methods, participation, and reliability. In WebSci. ACM, 234--243. Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Emily Pitler and Ani Nenkova. 2008. Revisiting readability: A unified framework for predicting text quality. In EMNLP. Association for Computational Linguistics, 186--195. Google ScholarGoogle ScholarDigital LibraryDigital Library
  32. Presentacion Rivera-Reyes. 2015. Students' task interpretation and conceptual understanding in electronics laboratory work. (2015).Google ScholarGoogle Scholar
  33. Libby O Ruch and Rae R Newton. 1977. Sex characteristics, task clarity, and authority. Sex Roles 3, 5 (1977), 479--494.Google ScholarGoogle ScholarCross RefCross Ref
  34. Aaron D Shaw, John J Horton, and Daniel L Chen. 2011. Designing incentives for inexpert human raters. In CSCW. ACM, 275--284. Google ScholarGoogle ScholarDigital LibraryDigital Library
  35. John Sweller and Paul Chandler. 1994. Why some material is difficult to learn. Cognition and instruction 12, 3 (1994), 185--233.Google ScholarGoogle Scholar
  36. Jie Yang, Claudia Hauff, Alessandro Bozzon, and Geert-Jan Houben. 2014. Asking the right question in collaborative q&a systems. In Hypertext. ACM, 179--189. Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Jie Yang, Judith Redi, Gianluca Demartini, and Alessandro Bozzon. 2016. Modeling task complexity in crowdsourcing. In HCOMP. AAAI, 249--258Google ScholarGoogle Scholar

Index Terms

  1. Clarity is a Worthwhile Quality: On the Role of Task Clarity in Microtask Crowdsourcing

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      HT '17: Proceedings of the 28th ACM Conference on Hypertext and Social Media
      July 2017
      336 pages
      ISBN:9781450347082
      DOI:10.1145/3078714

      Copyright © 2017 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 4 July 2017

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article

      Acceptance Rates

      HT '17 Paper Acceptance Rate19of69submissions,28%Overall Acceptance Rate378of1,158submissions,33%

      Upcoming Conference

      HT '24
      35th ACM Conference on Hypertext and Social Media
      September 10 - 13, 2024
      Poznan , Poland

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader