IDEAS home Printed from https://ideas.repec.org/a/inm/oropre/v62y2014i1p1-24.html
   My bibliography  Save this article

Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems

Author

Listed:
  • David R. Karger

    (Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139)

  • Sewoong Oh

    (Department of Industrial and Enterprise Systems Engineering, University of Illinois at Urbana--Champaign, Urbana, Illinois 61801)

  • Devavrat Shah

    (Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, Massachusetts 02139)

Abstract

Crowdsourcing systems, in which numerous tasks are electronically distributed to numerous “information pieceworkers,” have emerged as an effective paradigm for human-powered solving of large-scale problems in domains such as image classification, data entry, optical character recognition, recommendation, and proofreading. Because these low-paid workers can be unreliable, nearly all such systems must devise schemes to increase confidence in their answers, typically by assigning each task multiple times and combining the answers in an appropriate manner, e.g., majority voting.In this paper, we consider a general model of such crowdsourcing tasks and pose the problem of minimizing the total price (i.e., number of task assignments) that must be paid to achieve a target overall reliability. We give a new algorithm for deciding which tasks to assign to which workers and for inferring correct answers from the workers' answers. We show that our algorithm, inspired by belief propagation and low-rank matrix approximation, significantly outperforms majority voting and, in fact, is optimal through comparison to an oracle that knows the reliability of every worker. Further, we compare our approach with a more general class of algorithms that can dynamically assign tasks. By adaptively deciding which questions to ask to the next set of arriving workers, one might hope to reduce uncertainty more efficiently. We show that, perhaps surprisingly, the minimum price necessary to achieve a target reliability scales in the same manner under both adaptive and nonadaptive scenarios. Hence, our nonadaptive approach is order optimal under both scenarios. This strongly relies on the fact that workers are fleeting and cannot be exploited. Therefore, architecturally, our results suggest that building a reliable worker-reputation system is essential to fully harnessing the potential of adaptive designs.

Suggested Citation

  • David R. Karger & Sewoong Oh & Devavrat Shah, 2014. "Budget-Optimal Task Allocation for Reliable Crowdsourcing Systems," Operations Research, INFORMS, vol. 62(1), pages 1-24, February.
  • Handle: RePEc:inm:oropre:v:62:y:2014:i:1:p:1-24
    DOI: 10.1287/opre.2013.1235
    as

    Download full text from publisher

    File URL: http://dx.doi.org/10.1287/opre.2013.1235
    Download Restriction: no

    File URL: https://libkey.io/10.1287/opre.2013.1235?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Wenjie Wang & Lei Xie, 2022. "Optimal pricing of crowdsourcing logistics services with social delivery capacity," Journal of Combinatorial Optimization, Springer, vol. 43(5), pages 1447-1469, July.
    2. Vahideh Manshadi & Scott Rodilitz, 2022. "Online Policies for Efficient Volunteer Crowdsourcing," Management Science, INFORMS, vol. 68(9), pages 6572-6590, September.
    3. Tomer Geva & Maytal Saar‐Tsechansky, 2021. "Who Is a Better Decision Maker? Data‐Driven Expert Ranking Under Unobserved Quality," Production and Operations Management, Production and Operations Management Society, vol. 30(1), pages 127-144, January.
    4. Steve Alpern & Bo Chen, 2017. "Who should cast the casting vote? Using sequential voting to amalgamate information," Theory and Decision, Springer, vol. 83(2), pages 259-282, August.
    5. Banu Kabakulak & Z. Caner Taşkın & Ali Emre Pusane, 2021. "A branch-cut-and-price algorithm for optimal decoding in digital communication systems," Journal of Global Optimization, Springer, vol. 81(3), pages 805-834, November.
    6. Wang Chi Cheung & Will Ma & David Simchi-Levi & Xinshang Wang, 2022. "Inventory Balancing with Online Learning," Management Science, INFORMS, vol. 68(3), pages 1776-1807, March.
    7. Junming Yin & Jerry Luo & Susan A. Brown, 2021. "Learning from Crowdsourced Multi-labeling: A Variational Bayesian Approach," Information Systems Research, INFORMS, vol. 32(3), pages 752-773, September.
    8. Naonori Kakimura & Donghao Zhu, 2021. "Dynamic Bipartite Matching Market with Arrivals and Departures," Papers 2110.10824, arXiv.org.
    9. Yin, Xicheng & Wang, Hongwei & Wang, Wei & Zhu, Kevin, 2020. "Task recommendation in crowdsourcing systems: A bibliometric analysis," Technology in Society, Elsevier, vol. 63(C).
    10. Zhanwen Shi & Erbao Cao & Kai Nie, 2023. "Capacity pooling games in crowdsourcing services," Electronic Commerce Research, Springer, vol. 23(2), pages 1007-1047, June.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:inm:oropre:v:62:y:2014:i:1:p:1-24. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Chris Asher (email available below). General contact details of provider: https://edirc.repec.org/data/inforea.html .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.