A dataset containing the skeleton data extracted from videos of people performing
six different actions, recorded by a humanoid robot. Each class has several instances with
50 values (25 XY body keypoints extracted using OpenPose https://github.com/CMU-Perceptual-Computing-Lab/openpose)
per frame of the video (92 frames).
Usage
AR.data
Arguments
Format
A list of 6 different actions:
come
Gesture for telling the robot to come to you. 46 instances of [50x92] matrices.
five
Gesture of 'high five'. 45 instances of [50x92] matrices.
handshake
Gesture of handshaking with the robot. 45 instances of [50x92] matrices.
hello
Gesture for telling hello to the robot. 44 instances of [50x92] matrices.
ignore
Ignore the robot, pass by. 46 instances of [50x92] matrices.
look_at
Stare at the robot in front of it. 46 instances of [50x92] matrices.