Abstract
Crowdsourcing is a distributed problem-solving paradigm. Service oriented crowdsourcing paradigm involves both consumers and service providers. A consumer requests for a service (task); a provider provides that service (does that task); and the providers are paid by consumers for the service as per their satisfaction. The challenge is to select a service provider from a list of providers which can provide maximum satisfaction to the consumer for that service. This work outlines an architectural model using SLURM tool for efficient management of crowd. At the center of this work, we proposed a novel idea of adaptive task scheduling which is based on the customer satisfaction feedbacks. Our approach improves efficiency, and decreases the cost of service to consumers. Experimental results demonstrate the viability of our approach.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
J. Howe The rise of crowdsourcing (June 2006)
Zhou, T.C., Ma, H., King, I., Lyu, M.R.: TagRec: Leveraging Tagging Wisdom for Recommendation. In: IEEE Tran. On Computational Science and Engineering, CSE 2009, August 29-31, vol. 4, pp. 194–199 (2009)
Von Ahn, L., Dabbish, L.: Labeling Images with a Computer Game. In: CHI 2004: Proc. SIGCHI Conf. Human Factors in Computing Systems, pp. 319–326 (2004)
Yuen, M.-C., King, I., Leun, K.-S.: Task Matching in Crowdsourcing. In: IEEE International Conference on Cyber, Physical and Social Computing, vol. 4, pp. 409–412 (October 2011)
Yuen, M.-C., King, I., Kwong-Sak, L.: A Survey of Crowdsourcing Systems. In: IEEE International Conference on Social Computing (Socialcom), vol. 3, pp. 766–773 (October 2011)
Hirth, M., Hossfeld, T., Tran-Gia, P.: Anatomy of a Crowdsourcing Platform - Using the Example of Microworkers.com. In: IEEE International Conference on Innovative Mobile and Internet Services in Ubiquitous Computing (IMIS), vol. 5, pp. 322–329 (June 2011)
Zhang, H., Horvitz, E., Miller, R.C., Parkes, D.C.: Crowdsourcing General Computation. In: ACM CHI 2011 Workshop on Crowdsourcing and Human Computation (January 2011)
Psaier, H., Skopik, F., Schall, D., Dustdar, S.: Resource and Agreement Management in Dynamic Crowdcomputing Environments. In: IEEE Enterprise Distributed Object Computing Conference (EDOC), vol. 15, pp. 193–202 (August 2011)
Raykar, V.C., Yu, S.: An Entropic Score to Rank Annotators for Crowdsourced Labeling Tasks. In: IEEE Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG), vol. 3, pp. 29–32 (December 2011)
Hirth, M., Hossfeld, T., Tran-Gia, P.: Cost-Optimal Validation Mechanisms and Cheat-Detection for Crowdsourcing Platforms. In: Workshop on Future Internet and Next Generation Networks, Seoul, Korea (June 2011)
Mturk website, http://www.mturk.com
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Nunia, V., Kakadiya, B., Hota, C., Rajarajan, M. (2013). Adaptive Task Scheduling in Service Oriented Crowd Using SLURM. In: Hota, C., Srimani, P.K. (eds) Distributed Computing and Internet Technology. ICDCIT 2013. Lecture Notes in Computer Science, vol 7753. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-36071-8_29
Download citation
DOI: https://doi.org/10.1007/978-3-642-36071-8_29
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-36070-1
Online ISBN: 978-3-642-36071-8
eBook Packages: Computer ScienceComputer Science (R0)