How2Sign

A Large-scale Multimodal Dataset for Continuous American Sign Language

Dataset Information
Modalities
Texts, 3D, RGB-D, RGB Video
Languages
American Sign Language
Introduced
2020
Homepage

Overview

The How2Sign is a multimodal and multiview continuous American Sign Language (ASL) dataset consisting of a parallel corpus of more than 80 hours of sign language videos and a set of corresponding modalities including speech, English transcripts, and depth. A three-hour subset was further recorded in the Panoptic studio enabling detailed 3D pose estimation.

Variants: How2Sign

Associated Benchmarks

This dataset is used in 3 benchmarks:

Recent Benchmark Submissions

Task Model Paper Date
Sign Language Translation Unknown Sign Language Translation from Instructional … 2023-04-13
Video Inpainting INR-V INR-V: A Continuous Representation Space … 2022-10-29
Video Generation INR-V INR-V: A Continuous Representation Space … 2022-10-29

Research Papers

Recent papers with results on this dataset: