Abstract
A popular alternative non-conventional model for software development is crowdsourcing, which aims at decomposing software project into tasks that assigns them to individual stakeholders, through an open call for participation. A major challenge is to ensure community participation in developing high-quality solutions by each individual stakeholder. It is of high importance to be aware of the skills that can be acquired by the crowd, especially in cases of constantly evolving development environments, such as the JavaScript programming language and its applications. In the current paper, we aim at exploring trends in crowdsourcing JavaScript small tasks as an attempt to unveil a) popularity as the core technological skills and the functionalities that are more frequently crowdsourced, b) success as the relationship between the technological skills and the functionalities crowdsourced, and d) the monetary reward differences between these technological skills and the functionalities relationships. We have analyzed contest data collected from the Bountify crowdsourcing platform, resulting that while JavaScript small task development focuses on multiple technologies, frameworks, and libraries, that frequently overlap or complement each other, popularity, success, and monetary reward of the latter, in most cases are not associated.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Alt, F., Shirazi, A., Schmidt, A., Kramer, U., Nawaz, Z.: Location-based crowdsourcing: extending crowdsourcing to the real world. In: 6th Nordic Conf. on Human-Computer Interaction, Association for Computing Machinery, NY, USA, pp. 13–22 (2010)
Archak, N.: Money, glory and cheap talk: Analyzing strategic behavior of contestants in simultaneous crowdsourcing contests on TopCoder.com. In: 19th International Conference on WWW, NY, USA, pp. 21–30 (2010)
Bibi, S., Zozas, I., Ampatzoglou, A., Sarigiannidis, P.G., Kalampokis, G., Stamelos, I.: Crowdsourcing in software development. IEEE Access 8, 58094–58117 (2020)
Boutsis, I., Kalogeraki, V.: On task assignment for real-time reliable crowdsourcing. In: IEEE 34th International Conference on Distributed Computing Systems, Spain, pp. 1–10 (2014)
Chatzimparmpas, A., Bibi, S., Zozas, I., Kerren, A.: Analyzing the evolution of javascript applications. In: 14th International Conference on Evaluation of Novel Approaches to Software Engineering, vol. 1, pp. 359–366 (2019)
Delcev, S., Draskovic, D.: Modern javascript frameworks: a survey study. In: Zooming Innovation in Consumer Technologies Conference, Serbia, pp. 106–109 (2018)
Difallah, D., Catasta, M., Demartini, G., Ipeirotis, P., Cudré-Mauroux, P.: The dynamics of micro-task crowdsourcing: the case of Amazon MTurk. In: 24th International Conference on WWW, Switzerland, pp. 238–247 (2015)
Gude, S., Hafiz, M., Wirfs-Brock, A.: JavaScript: the used parts. In: IEEE 38th Annual Computer Software and Applications Conference, Sweden, pp. 466–475, (2014)
Guittard, C., Schenk, E., Burger-, T.: Crowdsourcing and the evolution of a business ecosystem. In: Garrigos, F.J., Gil-, I., Estelles, S. (eds.) Advances in Crowdsourcing, pp. 49–62. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-18341-1_4
Hetmank, L.: Components and functions of crowdsourcing systems – a systematic literature review. In: Wirtschaftsinformatik Proceedings, vol. 4 (2013)
1061–1998: IEEE Standard for a Software Quality Metrics Methodology, IEEE Standards, IEEE Computer Society, 31 December 1998 (reaf. 9 December 2009), (Accessed 21 July 2022)
Kittur, A., Chi, E., Suh, B.: Crowdsourcing user studies with Mechanical Turk. In: Conference on Human Factors in Computing Systems, NY, USA, pp. 453–456 (2008)
Kittur, A., et al.: The future of crowd work. Conference on Computer supported cooperative work, pp. 1301–1318. Association for Computing Machinery, NY, USA (2013)
Lakhani, K., Garvin, D., Lonstein, E.: TopCoder(A): Developing software through crowdsourcing. Harvard Business School Case (2010)
LaToza, T., Van der Hoek, A.: Crowdsourcing in software engineering: models, motivations, and challenges, IEEE Softw. 33(1), 74–80 (2016)
Machado, M., Zanatta, A., Marczack, S., Prikladnicki, R.: The good, the bad and the ugly: an onboard journey in software crowdsourcing competitive model. In: 4th International Workshop on Crowd Sourcing in Software Engineering, Buenos Aires, Argentina, pp. 2–8 (2017)
Mao, K., Capra, L., Harman, M., Jia, Y.: A survey of the use of crowdsourcing in software engineering. J. Syst. Softw. 126, 57–84 (2017)
Meldrum, S., Licorish, S., Savarimuthu, B. Crowdsourced knowledge on stack overflow. In: 21st International Conference on Evaluation and Assessment in Software Engineering. Association for Computing Machinery, NY, USA, pp. 180–185 (2017)
Papoutsoglou, M., Mittas, N., Angelis, L.: Mining people analytics from stackoverflow job advertisements. In: 43rd Euromicro Conference on Software Engineering and Advanced Applications, Austria, pp. 108–115 (2017)
Rauschmayer, A.: The Past, Present, and Future of JavaScript. O'Reilly Media, Inc. (2012)
Runeson, P., Höst, M.: Guidelines for conducting and reporting case study research in software engineering. Empir. Softw. Eng. 14, 131 (2009)
Sharma R., Sood M. Cloud SaaS: Models and Transformation. Advances in Digital Image Processing and Information Technology. CCIS, vol 205. Springer (2011). https://doi.org/10.1007/978-3-642-24055-3
Snook, S., Gorsuch, R.: Principal component analysis versus common factor analysis: A Monte Carlo study. Psychol. Bull. 106, 148–154 (1989)
Sun, K., Ryu, S.: Analysis of javaScript programs: challenges and research trends. ACM Comput. Surv. 50(4), Article 59, 34 (2017)
Tong, Y., Chen, L., Zhou, Z., Jagadish, H.V., Shou, L., Lv, W.: SLADE: A smart large-scale task decomposer in crowdsourcing. In: 35th IEEE International Conference on Data Engineering, ICDE Macau, China, pp. 2133–2134 (2019)
Wang, H., Wang, Y., Wang, J.: A participant recruitment framework for crowdsourcing-based software requirement acquisition. In: 9th IEEE International Conference on Global Software Engineering, pp. 65–73 (2014)
Weidema, E., López, C., Nayebaziz, S., Spanghero, G., Van der Hoek, A.: Toward microtask crowdsourcing software design work. In: 3rd International Workshop on Crowd Sourcing in Software Engineering, NY, USA, vol. 2016, pp. 41–44 (2016)
Yuen, M., King, I., Leung, K.: A survey of crowdsourcing systems. In: 3rd International Conference on Privacy, Security, Risk & Trust, MA, USA, pp. 766–773 (2011)
Zanatta, A., Machado, L., Steinmacher, I.: Competence, collaboration, and time management: barriers and recommendations for crowdworkers. In: 5th International Workshop on Crowd Sourcing in Software Engineering (CSI-SE), Gothenburg, Sweden, 9–16, (2018)
Zozas, I., Anagnostou, I., Bibi, S.: Trends on crowdsourcing javascript small tasks. ENASE 2022, 85–94 (2022)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Zozas, I., Anagnostou, I., Bibi, S. (2023). Identify Javascript Trends in Crowdsourcing Small Tasks. In: Kaindl, H., Mannion, M., Maciaszek, L.A. (eds) Evaluation of Novel Approaches to Software Engineering. ENASE 2022. Communications in Computer and Information Science, vol 1829. Springer, Cham. https://doi.org/10.1007/978-3-031-36597-3_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-36597-3_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-36596-6
Online ISBN: 978-3-031-36597-3
eBook Packages: Computer ScienceComputer Science (R0)