This point cloud dataset is captured by 10 "synchronized" Kinects (named the Kinoptic Studio) installed in the Panoptic Studio
There is no way to perfectly synchronize multiple kinects, but we accurately aligned them by some hardware modifications for time-stamping.
The 3D point cloud is generated by merging the depth maps from the multiple Kinects captured within a time interval (+-15msec). This means that if the human motion is fast, there exist misalignment.
The kinect data is captured with other 500+ RGB cameras, and they are sharing timespace and 3D world coordinate. So this point cloud output can be used with the output of RGB cameras (e.g. RGB videos and 3D skeletons).
For more hardware details, please see our tutorial page.
Data Description
The basic data types are synchronized RGB videos from 10 kinects with corresponding 10 depth files.
You can easily generate point cloud using our KinopticStudio Toolbox in our github page.
Frame indexing: the frame index of the output rgb image frames and 3D point clouds is defined by the HD camera frame indexing of the panoptic studio. Note that Panoptic Studio have two different frame indexing, VGA and HD.