Recognizing kitchen tools and ingredients from shifting, shaky angles.
Often includes synchronized gaze data (where the person is looking) Content and Activity
If you tell me more about your specific project, I can provide: for this specific timestamp (if available) Code snippets for loading GTEA Gaze+ videos in Python Related research papers that utilize the Group 4 dataset g4_01136.mp4
In this specific sequence, a subject is filmed in a natural kitchen setting performing a "recipe-driven" task.
Researchers use "g4_01136.mp4" and similar clips to train and test algorithms in several key areas: Typically involves preparing a specific meal, such as
Modeling how a person’s eyes move toward an object before their hands touch it.
Typically involves preparing a specific meal, such as making a sandwich, salad, or tea. Typically involves preparing a specific meal
🎥 This video is often cited in papers involving or Transformers designed for video understanding. It serves as a "real-world" challenge because of motion blur, hand occlusions, and the visual complexity of a cluttered kitchen.