Atomic Visual Actions
AVA is a project that provides audiovisual annotations of video for improving our understanding of human activity. Each of the video clips has been exhaustively annotated by human annotators, and together they represent a rich variety of scenes, recording conditions, and expressions of human activity. There are annotations for:
Variants: AVA v2.1, AVA-ActiveSpeaker, AVA-LAEO, AVA-Speech, AVA v2.2, AVA-Kinetics
This dataset is used in 1 benchmark:
Task | Model | Paper | Date |
---|---|---|---|
Node Classification | ASDNet [ASDNet_ICCV2021] | Learning Long-Term Spatial-Temporal Graphs for … | 2022-07-15 |
Node Classification | TalkNet [tao2021someone] | Learning Long-Term Spatial-Temporal Graphs for … | 2022-07-15 |
Node Classification | UniCon [zhang2021unicon] | Learning Long-Term Spatial-Temporal Graphs for … | 2022-07-15 |
Node Classification | MAAS-TAN [MAAS2021] | Learning Long-Term Spatial-Temporal Graphs for … | 2022-07-15 |
Recent papers with results on this dataset: