• DocumentCode
    710102
  • Title

    Comprehensive and reliable crowd assessment algorithms

  • Author

    Joglekar, Manas ; Garcia-Molina, Hector ; Parameswaran, Aditya

  • Author_Institution
    Comput. Sci. Dept., Stanford Univ., Stanford, CA, USA
  • fYear
    2015
  • fDate
    13-17 April 2015
  • Firstpage
    195
  • Lastpage
    206
  • Abstract
    Evaluating workers is a critical aspect of any crowdsourcing system. In this paper, we devise techniques for evaluating workers by finding confidence intervals on their error rates. Unlike prior work, we focus on “conciseness”-that is, giving as tight a confidence interval as possible. Conciseness is of utmost importance because it allows us to be sure that we have the best guarantee possible on worker error rate. Also unlike prior work, we provide techniques that work under very general scenarios, such as when not all workers have attempted every task (a fairly common scenario in practice), when tasks have non-boolean responses, and when workers have different biases for positive and negative tasks. We demonstrate conciseness as well as accuracy of our confidence intervals by testing them on a variety of conditions and multiple real-world datasets.
  • Keywords
    behavioural sciences computing; personnel; comprehensive crowd assessment algorithms; conciseness; confidence intervals; multiple real-world datasets; nonBoolean responses; reliable crowd assessment algorithms; worker error rate; Accuracy; Crowdsourcing; Error analysis; Gold; Random variables; Reliability; Standards;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Data Engineering (ICDE), 2015 IEEE 31st International Conference on
  • Conference_Location
    Seoul
  • Type

    conf

  • DOI
    10.1109/ICDE.2015.7113284
  • Filename
    7113284