An experimental characterization of workers' behavior and accuracy in crowdsourced tasks.
Crowdsourcing systems are evolving into a powerful tool of choice to deal with repetitive or lengthy human-based tasks. Prominent among those is Amazon Mechanical Turk, in which Human Intelligence Tasks, are posted by requesters, and afterwards selected and executed by subscribed (human) workers in...
Guardado en:
Autores principales: | Evgenia Christoforou, Antonio Fernández Anta, Angel Sánchez |
---|---|
Formato: | article |
Lenguaje: | EN |
Publicado: |
Public Library of Science (PLoS)
2021
|
Materias: | |
Acceso en línea: | https://doaj.org/article/974753e867284d2383e4f4cc67f2f518 |
Etiquetas: |
Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
|
Ejemplares similares
-
Mitigating Herding in Hierarchical Crowdsourcing Networks
por: Han Yu, et al.
Publicado: (2016) -
Crowdsourcing novel childhood predictors of adult obesity.
por: Kirsten E Bevelander, et al.
Publicado: (2014) -
Crowdsourced analysis of fungal growth and branching on microfluidic platforms.
por: Alex Hopke, et al.
Publicado: (2021) -
openSNP--a crowdsourced web resource for personal genomics.
por: Bastian Greshake, et al.
Publicado: (2014) -
Atrial fibrillation detection in outpatient electrocardiogram monitoring: An algorithmic crowdsourcing approach.
por: Ali Bahrami Rad, et al.
Publicado: (2021)