Skip to main content

ORIGINAL RESEARCH article

Front. Organ. Psychol.

Sec. Performance and Development

Volume 3 - 2025 | doi: 10.3389/forgp.2025.1500016

Influence of Deliverable Evaluation Feedback and Additional Reward on Worker's Motivation in Crowdsourcing Services

Provisionally accepted
Yoshinori Hijikata Yoshinori Hijikata 1,2*Hikari Ishizaki Hikari Ishizaki 1
  • 1 Kwansei Gakuin University, Nishinomiya, Japan
  • 2 University of Hyogo, Kobe, Hyōgo, Japan

The final, formatted version of the article will be published soon.

    To create training data for AI systems, it is necessary to manually assign correct labels to a large number of objects; this task is often performed by crowdsourcing. This task is usually divided into a certain number of smaller and more manageable segments, and workers work on them one after the other. In this study, assuming the above task, we investigated whether the deliverable evaluation feedback and provision of additional rewards contribute to the improvement of workers' motivation, that is, the persistence of the tasks and performance.We conducted a user experiment on a real crowdsourcing service platform. This provided first and second round of tasks, which ask workers input correct labels to a flower species. We developed an experimental system that assessed the work products of the first-round task performed by a worker and presented the results to the worker. 645 workers participated in this experiment. They were divided into high and low performing groups according to their first-round scores (correct answer ratio). The workers' performance and task continuation ratio under the high and low performance group and with and without evaluation feedback and additional rewards were compared. We found that the presentation of deliverable evaluations increased the task continuation rate of high-quality workers, but did not contribute to an increase in the task performance (correct answer rate) for either type of worker. The providing additional rewards reduced workers' task continuation rate, and the amount of reduction was larger for low-quality workers than that for highquality workers. However, it largely increased the low-quality worker's task performance. Although not statistically significant, the low-quality worker's task performance of the second round was highest for those who were shown both feedback and additional rewards. In conclusion, it is better to offer both feedback and additional rewards when the quality of the deliverables is a priority, and to give only feedback when the quantity of deliverables is a priority.

    Keywords: crowdsourcing, deliverable assessment, Evaluation feedback, additional reward, Worker motivation, task continuation, task performance

    Received: 29 Sep 2024; Accepted: 18 Mar 2025.

    Copyright: © 2025 Hijikata and Ishizaki. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence: Yoshinori Hijikata, Kwansei Gakuin University, Nishinomiya, Japan

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

    People also looked at

    Research integrity at Frontiers

    Man ultramarathon runner in the mountains he trains at sunset

    94% of researchers rate our articles as excellent or good

    Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


    Find out more