top of page

G4_01136.mp4 -

🎥 This video is often cited in papers involving or Transformers designed for video understanding. It serves as a "real-world" challenge because of motion blur, hand occlusions, and the visual complexity of a cluttered kitchen.

Modeling how a person’s eyes move toward an object before their hands touch it.

Recognizing kitchen tools and ingredients from shifting, shaky angles. g4_01136.mp4

A consistent kitchen laboratory setup used across the "g4" (Group 4) subset of the data. Technical Significance

In this specific sequence, a subject is filmed in a natural kitchen setting performing a "recipe-driven" task. 🎥 This video is often cited in papers

If you tell me more about your specific project, I can provide: for this specific timestamp (if available) Code snippets for loading GTEA Gaze+ videos in Python Related research papers that utilize the Group 4 dataset

Understanding the logical sequence of steps required to complete a complex task. Usage in AI Benchmarking If you tell me more about your specific

Often includes synchronized gaze data (where the person is looking) Content and Activity

bottom of page