The slide presents a four-step workflow for gaze tracking methodology. It covers eye/pupil detection with YOLOv5, 3D gaze vector estimation using CNN, 2D point-of-regard prediction via polynomial regression, and output smoothing with a Kalman filter.
Methodology
{ "headers": [ "Step", "Description", "Technology" ], "rows": [ [ "1. Eye/Pupil Detection", "Detect eyes and pupils from real-time video input", "YOLOv5" ], [ "2. Gaze Vector Estimation", "Compute 3D gaze direction vector from eye images", "CNN" ], [ "3. Point-of-Regard Prediction", "Map gaze vector to 2D screen coordinates", "Polynomial Regression" ], [ "4. Smoothing & Output", "Apply temporal filtering for stable gaze tracking", "Kalman Filter" ] ] }
Source: EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression
Speaker Notes
Walk through the step-by-step workflow: Start with robust eye detection using YOLOv5, estimate gaze vectors with CNN, predict point-of-regard via polynomial regression, and smooth outputs with Kalman filter for precise, real-time HCI.