RoboCrowd: Scaling Robot Data Collection through Crowdsourcing

Stanford University

Abstract

In recent years, imitation learning from large-scale human demonstrations has emerged as a promising paradigm for training robot policies. However, the burden of collecting large quantities of human demonstrations is significant in terms of collection time and the need for access to expert operators. We introduce a new data collection paradigm, RoboCrowd, which distributes the workload by utilizing crowdsourcing principles and incentive design. RoboCrowd helps enable scalable data collection and facilitates more efficient learning of robot policies. We build RoboCrowd on top of ALOHA—a bimanual platform that supports data collection via puppeteering—to explore the design space for crowdsourcing in-person demonstrations in a public environment. We propose three classes of incentive mechanisms to appeal to users' varying sources of motivation for interacting with the system: material rewards, intrinsic interest, and social comparison. We instantiate these incentives through tasks that include physical rewards, engaging or challenging manipulations, as well as gamification elements such as a leaderboard. We conduct a large-scale, two-week field experiment in which the platform is situated in a university cafe. We observe significant engagement with the system—over 200 individuals independently volunteered to provide a total of over 800 interaction episodes. Our findings validate the proposed incentives as mechanisms for shaping users' data quantity and quality. Further, we demonstrate that the crowdsourced data can serve as useful pre-training data for policies fine-tuned on expert demonstrations—boosting performance up to 20% compared to when this data is not available. These results suggest the potential for RoboCrowd to reduce the burden of robot data collection by carefully implementing crowdsourcing and incentive design principles.

RoboCrowd in a cafe setting

We deployed RoboCrowd in a university cafe, where over 200 individuals provided data with the system over two weeks.


Dataset Samples

Sample Expert Demonstrations

Select one of the tasks below to view an expert trajectory for that task.

Top
Left
Right

Sample Crowdsourced Interactions

Select one of the examples below to view a sample crowdsourced interaction from the dataset (which consists of over 800 interaction episodes).

Top
Left
Right

BibTeX


@article{mirchandani2024robocrowd,
    title   = {RoboCrowd: Scaling Robot Data Collection through Crowdsourcing},
    author  = {Suvir Mirchandani and David D. Yuan and Kaylee Burns and Md Sazzad Islam and Tony Z. Zhao and Chelsea Finn and Dorsa Sadigh},
    journal = {arXiv},
    year    = {2024},
}