JHMDB

Joint-annotated Human Motion Data Base

Dataset Information
Modalities
Videos
Introduced
2013
License
Unknown
Homepage

Overview

JHMDB is an action recognition dataset that consists of 960 video sequences belonging to 21 actions. It is a subset of the larger HMDB51 dataset collected from digitized movies and YouTube videos. The dataset contains video and annotation for puppet flow per frame (approximated optimal flow on the person), puppet mask per frame, joint positions per frame, action label per clip and meta label per clip (camera motion, visible body parts, camera viewpoint, number of people, video quality).

Source: Unsupervised Deep Metric Learning via Orthogonality based Probabilistic Loss
Image Source: https://arxiv.org/pdf/1712.06316.pdf

Variants: JHMDB (2D poses only), J-HMDB, J-HMBD Early Action, JHMDB Pose Tracking, J-HMDB-21, JHMDB

Associated Benchmarks

This dataset is used in 1 benchmark:

Recent Benchmark Submissions

Task Model Paper Date
Open Vocabulary Action Detection SiA Scaling Open-Vocabulary Action Detection 2025-04-04

Research Papers

Recent papers with results on this dataset: