EgoExoLearn

Dataset Information
Introduced
2024
License
Unknown
Homepage

Overview

EgoExoLearn is a fascinating dataset designed to bridge the gap between egocentric and exocentric views of procedural activities.

  1. What Is EgoExoLearn? EgoExoLearn is a large-scale dataset that emulates how humans learn by observing others. It focuses on the process of asynchronous demonstration following. Participants in the dataset record egocentric videos as they perform tasks. These videos are guided by exocentric-view demonstration videos. In simpler terms, imagine someone watching a demonstration video (from an external perspective) and then replicating the same task while recording their own point-of-view video.

  2. Dataset Details:
    EgoExoLearn dataset spans 120 hours and covers scenarios from daily life and specialized laboratories. It contains:

    • Egocentric videos: These are recorded by individuals executing tasks.
    • Demonstration videos: These show the same tasks from an external viewpoint.
    • Gaze data: High-quality gaze information accompanies the videos.
    • Multimodal annotations: Detailed annotations provide context and insights.
  3. Applications and Benchmarks:
    The EgoExoLearn dataset serves as a playground for modeling the human ability and thus provides a playground to bridge asynchronous procedural actions from different viewpoints.
    It enables new benchmarks such as:

    • Cross-view association: Linking actions observed from different perspectives.
    • Cross-view action planning: Anticipating and planning actions based on both ego and exo views.
    • Cross-view referenced skill assessment: Evaluating skills across viewpoints. Using Exo-view demonstrations as guidance for better ego-view skill assessment.
  4. Why Is It Important?

    • Understanding how we map others' activities into our own point of view is a fundamental human skill.
    • EgoExoLearn paves the way for creating AI agents capable of seamlessly learning by observing humans in the real world.

(1) EgoExoLearn: A Dataset for Bridging Asynchronous Ego- and Exo-centric .... https://arxiv.org/html/2403.16182v1.
(2) EgoExoLearn: A Dataset for Bridging Asynchronous Ego- and Exo-centric .... https://egoexolearn.github.io/.
(3) EgoExoLearn: A Dataset for Bridging Asynchronous Ego- and Exo-centric .... https://arxiv.org/abs/2403.16182.
(4) EgoExoLearn: A Dataset for Bridging Asynchronous Ego- and Exo-centric .... https://allainews.com/item/egoexolearn-a-dataset-for-bridging-asynchronous-ego-and-exo-centric-view-of-procedural-activities-in-real-world-2024-03-26/.
(5) undefined. https://github.com/OpenGVLab/EgoExoLearn/.

Variants: EgoExoLearn

Associated Benchmarks

This dataset is used in 3 benchmarks:

Recent Benchmark Submissions

Task Model Paper Date
Action Quality Assessment RAAN+TL+Gaze EgoExoLearn: A Dataset for Bridging … 2024-03-24
Action Quality Assessment RAAN+TL EgoExoLearn: A Dataset for Bridging … 2024-03-24
Action Anticipation Action anticipation baseline (co-training, with gaze) EgoExoLearn: A Dataset for Bridging … 2024-03-24
Action Anticipation Action anticipation baseline (co-training, no gaze) EgoExoLearn: A Dataset for Bridging … 2024-03-24
Video Retrieval cross-view association baseline (gaze, val) EgoExoLearn: A Dataset for Bridging … 2024-03-24
Video Retrieval cross-view association baseline (no gaze, val) EgoExoLearn: A Dataset for Bridging … 2024-03-24

Research Papers

Recent papers with results on this dataset: