Egocentric Benchmarks
Action Recognition & Activity Understanding
- EPIC-KITCHENS: Unscripted cooking activities with object interactions, also involves EPIC Fields and EPIC-SOUNDS
- Ego4D: Large-scale daily activities (1,400+ hrs, 2.2M annotations), it also has EgoTracks as a part.
- EGTEA Gaze: Kitchen activities with gaze tracking
- EGTEA Gaze+: Another daatset with Gaze Tracking. There is GTEA Gaze, Gaze+, sub, EGTEA Gaze+
- Charades-Ego: Paired third-person and egocentric daily activities
- ASCC Activities of Daily Living: Recorded with a chest-mounted camera, showing activities of daily living such as cooking, cleaning, etc.
- EgoPet: From pet based egocentric view
- AEA: Everyday Activity dataset
- DataEgo: egocentric dataset composed of visual, gyroscope and accelerometer information
- EgoPER: Egocentric Procedural Task Videos
- WEAR: Wearable and Egocentric Activity Recognition
- IndustReal: Action recognition and assembly state detection
- EgoProceL: Dataset for procedural learning
Hand Interaction
- EgoHands: Hand detection and segmentation
- EgoDexter
- H2O: Human-human-object interactions
- EgoGesture: multi-modal large scale dataset for egocentric hand gesture recognition
- EgoChoir: estimate 3D human contact and obejct affordance from egocentric videos
- HOT3D: 3D hand and object tracking
- ARCTIC: Dexterous Bimanual Hand-Object Manipulation
- HOI4D: Category-Level Human-Object Interaction
Long-Term Video Understanding
- DailyLife: Long-term activity recognition (12 hours/day)
- EgoRoutine: Routine understanding over extended periods
Social Interaction
- EgoCom: Multi-person conversations
- EgoReID: Person Reidentification
- EgoHumans: Egocentric human 3D pose estimation and tracking
Navigation and Scene Understanding
- VizWiz-Ego: Assistive vision for visually impaired users
- Ego-Centric: Object interactions in daily environments
Gaze and Attention Prediction
- EHTask: Gaze behavior during manipulation tasks
Motion and Pose Estimation
- EgoCap: 3D body pose from head-mounted cameras
- xr-EgoPose: Egocentric 3D human pose estimation
- Nymeria: Human Pose Estimation
- Ego-Exo4D: Meticulously synchronized natural language datasets paired with videos
- EgoExoLearn: Asynchronous Ego- and Exo-centric View of Procedural Activities in Real World
- Assembly101: Multi-View Video Dataset for Understanding Procedural Activities
- Mo2Cap2: Egocentric estimation of 3D human body pose in a wide range of unconstrained everyday activities
- EgoWholeMocap: Egocentric whole-body motion capture using a single fisheye camera, which simultaneously estimates human body and hand motion
- SceneEgo: Scene-aware Egocentric 3D Human Pose Estimation
- EE3D: Monocular egocentric 3D human motion capture
- EgoBody3M: Accurate tracking of a user’s body pose while wearing a virtual reality (VR), augmented reality (AR) or mixed reality (MR)
- AssemblyHands: 3D Hand Pose Estimation
3D Human Motion Capture
- UnrealEgo: Egocentric 3D Human Motion Capture
- UnrealEgo2: Stereo egocentric 3D human pose estimation
- FPHA: RGB-D Videos and 3D Hand Pose Annotations
Object Detection
- TACO: Image dataset of waste in the wild
- HInt: Reconstructing Hands in 3D with Transformers
- ADL: detecting activities of daily living (ADL) in first-person camera views
Image Based Localization
- EgoCART: Indoor Image-Based Localization