Video debug information:

Loading video information...

InclusiveVidPose: First Video Pose Dataset for Limb Deficiencies

Data Collection

  • Why do we build our dataset from videos instead of curated images?
  • Data sourcing
  • Data filtering and curation

Our Keypoint Schema

  • Necessity of Specialized Human Body Keypoint Schema
  • Keypoint Schema Design

Data Annotation Procedure

  • Annotation Team Training
  • Bounding Box, Segmentation Mask and Tracking ID
  • Keypoint Annotation Process

Dataset Statistics

  • More details about dataset statistics

Why do we need InclusiveVidPose?

Approximately 40 million individuals worldwide live with limb deficiencies. These people could benefit greatly from accurate human pose estimation (HPE) in applications such as rehabilitation monitoring and health assessment. Regrettably, existing HPE systems trained on standard datasets fail to accommodate atypical anatomies and prosthetic occlusions, leaving this population unsupported. However, widely used benchmarks like MS COCO and MPII Human Pose include only able-bodied people with complete sets of keypoints. These datasets and the methods built on them assume that every keypoint of a presented individual exists, making no provision for missing or altered limbs. As demonstrated in the figure next, the MS COCO–trained ViTPose model produces significant errors when applied to images of individuals with limb deficiencies. Consequently, people with limb deficiencies are excluded from current benchmarks, and models trained on these datasets fail to generalize to their anatomies. To address this gap, we introduce InclusiveVidPose, the first video-based HPE dataset focused on individuals with limb deficiencies.

Why We Need InclusiveVidPose

What is InclusiveVidPose?

InclusiveVidPose Dataset, the first video-based HPE dataset specific for individuals with limb deficiencies. We collected 313 videos, totaling nearly 340k frames, and covering over 400 individuals with amputations, congenital limb differences, and prosthetic limbs. We introduce custom keypoints at each residual limb end to capture individual anatomical variations.

InclusiveVidPose Dataset Features

  • 313 high-quality video sequences
  • 345,081 annotated frames
  • 401 unique individuals with limb deficiencies
  • 25 keypoints with 8 residual-limb end
  • Subject segmentation mask
  • Subject bounding box
  • Subject tracking ID
Research Paper Cover

InclusiveVidPose: A Video-based Dataset for Human Pose Estimation of Individuals with Limb Deficiencies

Authors: Anonymous Authors

Published: Conference Name, Year

Dataset Examples

Dataset Examples

email@example.com

+123456789

Location