|Description (include details on usage, files and paper references)||JPL First-Person Interaction dataset (JPL-Interaction dataset) is composed of human activity videos taken from a first-person viewpoint. The dataset particularly aims to provide first-person videos of interaction-level activities, recording how things visually look from the perspective (i.e., viewpoint) of a person/robot participating in such physical interactions.
This first-person dataset contains videos of interactions between humans and the observer. We attached a GoPro2 camera to the head of our humanoid model, and asked human participants to interact with the humanoid by performing activities. In order to emulate the mobility of a real robot, we also placed wheels below the humanoid and made an operator to move the humanoid by pushing it from the behind.
There are 7 different types of activities in the dataset, including 4 positive (i.e., friendly) interactions with the observer, 1 neutral interaction, and 2 negative (i.e., hostile) interactions. "Shaking hands with the observer", "hugging the observer", "petting the observer", and "waving a hand to the observer" are the four friendly interactions. The neutral interaction is the situation where two persons have a conversation about the observer while occasionally pointing it. "Punching the observer" and "throwing objects to the observer" are the two negative interactions. Videos were recorded continuously during human activities where each video sequence contains 0 to 3 activities. The videos are in 320*240 resolution with 30 fps.
First-Person Activity Recognition: What Are They Doing to Me
CVPR 2013, M. S. Ryoo and L. Matthies