Joint-annotated Human Motion Data Base
JHMDB is an action recognition dataset that consists of 960 video sequences belonging to 21 actions. It is a subset of the larger HMDB51 dataset collected from digitized movies and YouTube videos. The dataset contains video and annotation for puppet flow per frame (approximated optimal flow on the person), puppet mask per frame, joint positions per frame, action label per clip and meta label per clip (camera motion, visible body parts, camera viewpoint, number of people, video quality).
Source: Unsupervised Deep Metric Learning via Orthogonality based Probabilistic Loss
Image Source: https://arxiv.org/pdf/1712.06316.pdf
Variants: JHMDB (2D poses only), J-HMDB, J-HMBD Early Action, JHMDB Pose Tracking, J-HMDB-21, JHMDB
This dataset is used in 1 benchmark:
Task | Model | Paper | Date |
---|---|---|---|
Open Vocabulary Action Detection | SiA | Scaling Open-Vocabulary Action Detection | 2025-04-04 |
Recent papers with results on this dataset: