EyeScribe: AI Eye-Tracking for Hands-Free HCI (42 chars)

Generated from prompt:

Create a presentation titled 'EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression'. Domain: Artificial Intelligence / Computer Vision / Human-Computer Interaction (HCI). Follow this slide structure: 1. Introduction - Introduction - Existing System - Problem Statement - Objectives 2. Literature Review - Literature Review - Base Paper - Literature Review - Additional Paper 3. Research Gap - Research Gap 4. Proposed System - Proposed System - Methodology - Data Used - Features - Tech Stack Use the abstract and project description provided to generate slide content. Include diagrams or icons where suitable, and maintain a modern AI/tech theme with blue and white tones.

Introduces EyeScribe, a low-cost vision-based HCI using DL gaze detection & geometric regression. Covers existing systems, literature, gaps, methodology, data, features, and tech stack for accurate, r

December 15, 202517 slides
Slide 1 of 17

Slide 1 - EyeScribe

This title slide presents "EyeScribe: A Vision-Based Assistive Human-Computer Interface." The subtitle describes its use of deep learning and geometric regression in AI, computer vision, and HCI, featuring an eye icon and neural net graphic.

EyeScribe: A Vision-Based Assistive Human-Computer Interface

Using Deep Learning and Geometric Regression in AI, Computer Vision & HCI (Include eye icon & neural net graphic)

Slide 1 - EyeScribe
Slide 2 of 17

Slide 2 - EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

This slide serves as the section header for "Section 1: Introduction (01)" in the EyeScribe project, titled "EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression." The subtitle outlines an overview of the project structure, existing systems, problem statement, and objectives.

EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

01

Section 1: Introduction

Overview of project structure, existing systems, problem statement, and objectives

Source: Project Description

Speaker Notes
Introduce the EyeScribe project in AI, Computer Vision, and HCI. This section covers: Introduction, Existing System, Problem Statement, and Objectives. Highlight the modern AI/tech theme with blue and white tones.
Slide 2 - EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression
Slide 3 of 17

Slide 3 - Introduction

This introduction slide outlines a hands-free human-computer interaction (HCI) system powered by eye-tracking technology for assistive use by disabled users. It leverages deep learning for accurate gaze detection and geometric regression for precise control.

Introduction

  • Enables hands-free HCI via eye-tracking technology
  • Uses Deep Learning for accurate gaze detection
  • Applies geometric regression for precise control
  • Targets assistive technology for disabled users

Source: Project Description

Speaker Notes
Include eye diagram. Highlight EyeScribe's innovation in hands-free HCI for assistive tech.
Slide 3 - Introduction
Slide 4 of 17

Slide 4 - Existing System

The slide outlines existing eye-tracking systems: the hardware-heavy, high-cost Tobii Eye Tracker and the low-accuracy, webcam-based WebGazer. It notes key limitations, including imprecision in varied lighting and the need for specialized setups or reduced precision.

Existing System

  • • Tobii Eye Tracker: Hardware-heavy, high cost
  • • WebGazer: Webcam-based, low accuracy
  • • Limitations: Imprecise in varied lighting
  • • Require specialized setup or sacrifice precision

Source: Project Description

Speaker Notes
Highlight current solutions like Tobii Eye Tracker (hardware-intensive, costly) and WebGazer (web-based, low accuracy). Emphasize limitations such as high costs and poor performance in varied lighting. Suggest comparison table icon for visual contrast.
Slide 4 - Existing System
Slide 5 of 17

Slide 5 - Problem Statement

This slide outlines key problems in vision-based HCI interfaces, including a lack of affordable, accurate options and pupil detection errors in varying lighting. It also highlights head movement interference, high real-time computational demands, and limited accessibility for assistive technologies.

Problem Statement

  • Lack of affordable, accurate vision-based HCI interfaces.
  • Pupil detection errors in varying lighting conditions.
  • Head movement interference degrades tracking reliability.
  • Real-time processing demands high computational efficiency.
  • Limited accessibility for assistive technologies.

Source: EyeScribe: A Vision-Based Assistive Human-Computer Interface

Speaker Notes
Highlight challenges in vision-based HCI: affordability, accuracy issues with pupil detection, head movements, and real-time needs. Suggest problem icons (e.g., warning symbols).
Slide 5 - Problem Statement
Slide 6 of 17

Slide 6 - Agenda

The agenda slide outlines a four-part presentation structure. It covers: 1) Introduction to overview, existing systems, problem, and objectives; 2) Literature review of base and relevant papers; 3) Identified research gaps; and 4) Proposed system with methodology, data, features, and tech stack.

Agenda

  1. 1. Introduction
  2. Overview, existing systems, problem statement, and objectives.

  3. 2. Literature Review
  4. Base paper and additional relevant papers.

  5. 3. Research Gap
  6. Key gaps identified in existing literature.

  7. 4. Proposed System

Methodology, data, features, and technology stack. Source: EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

Speaker Notes
Develop real-time eye-tracking system. Achieve >95% gaze accuracy. Integrate geometric regression for cursor control. Build accessible prototype.
Slide 6 - Agenda
Slide 7 of 17

Slide 7 - EyeScribe Presentation

This slide serves as the section header for "Literature Review" (section 02) in the EyeScribe Presentation. Its subtitle highlights key papers surveyed in vision-based assistive HCI and deep learning.

EyeScribe Presentation

02

Literature Review

Key papers surveyed in vision-based assistive HCI and deep learning

Source: AI/CV/HCI Domain

Speaker Notes
Introduce key papers surveyed in vision-based assistive interfaces, highlighting relevant deep learning and HCI advancements.
Slide 7 - EyeScribe Presentation
Slide 8 of 17

Slide 8 - Literature Review - Base Paper

The slide reviews Zhang et al.'s base paper, which uses CNNs for precise pupil/iris detection from eye images to map movements to gaze directions. It notes achievements like 4° accuracy in controlled settings, but limitations in handling head movement and varying lighting with static setups.

Literature Review - Base Paper

Deep Learning for Gaze Estimation (Zhang et al.)Achievements & Limitations
Employs Convolutional Neural Networks (CNN) for accurate pupil and iris detection from eye images. Extracts features to map eye movements to gaze directions effectively. (Paper icon)Achievements: 4° gaze estimation accuracy in controlled settings. Limitations: Relies on static setups; struggles with head movement and varying lighting. (Paper icon)

Source: Zhang et al., Deep Learning for Gaze Estimation

Speaker Notes
Highlight the foundational CNN approach for gaze estimation, its accuracy milestone, and key limitations in dynamic scenarios.
Slide 8 - Literature Review - Base Paper
Slide 9 of 17

Slide 9 - Literature Review - Additional Paper

The slide reviews Smith et al.'s paper on Geometric Regression in Eye-Tracking, using polynomial models for precise gaze mapping. It emphasizes enhanced robustness to head pose variations, improved real-world accuracy, and includes a graph of visual gaze estimation results.

Literature Review - Additional Paper

  • Geometric Regression in Eye-Tracking (Smith et al.)
  • Polynomial models for precise gaze mapping
  • Enhances robustness to head pose variations
  • Improves accuracy in real-world scenarios
  • Graph icon: Visual gaze estimation results

Source: Geometric Regression in Eye-Tracking (Smith et al.)

Speaker Notes
Highlight graph icon for polynomial gaze mapping visualization. Emphasize head pose robustness.
Slide 9 - Literature Review - Additional Paper
Slide 10 of 17

Slide 10 - EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

This section header slide introduces Section 3 titled "Research Gap" in the EyeScribe presentation. Its subtitle highlights identified opportunities in existing vision-based assistive interfaces for enhanced human-computer interaction.

EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

3

Research Gap

Identified opportunities in existing vision-based assistive interfaces for enhanced HCI

Source: Artificial Intelligence / Computer Vision / Human-Computer Interaction (HCI)

Speaker Notes
Emphasize identified opportunities in current literature, such as limitations in real-time accuracy, integration challenges, and lack of geometric regression for precise assistive control.
Slide 10 - EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression
Slide 11 of 17

Slide 11 - Research Gap

The slide identifies key research gaps in low-cost assistive HCI, including no integrated deep learning and geometric systems or real-time mobile deployment. It also highlights insufficient diverse datasets for training and challenging, non-user-friendly calibration processes.

Research Gap

  • No integrated DL + geometric system for low-cost assistive HCI.
  • Lack of real-time deployment on mobile devices.
  • Insufficient diverse datasets for robust training.
  • Challenging user calibration without ease-of-use.

Source: EyeScribe Presentation

Speaker Notes
No integrated DL+geometric system for low-cost assistive HCI. Key gaps: real-time mobile deployment, diverse datasets, user calibration ease. Include gap diagram.
Slide 11 - Research Gap
Slide 12 of 17

Slide 12 - EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

This section header slide introduces Section 04: Proposed System, under the title "EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression." The subtitle emphasizes "EyeScribe Architecture for Assistive Human-Computer Interaction."

EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

04

Proposed System

EyeScribe Architecture for Assistive Human-Computer Interaction

Source: AI / Computer Vision / HCI

Speaker Notes
Introduce the EyeScribe architecture, highlighting key components before diving into methodology, data, features, and tech stack.
Slide 12 - EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression
Slide 13 of 17

Slide 13 - Proposed System

The proposed system slide outlines a pipeline starting with a camera capturing real-time eye video feed. A DL Gaze Detector estimates gaze direction, refined by a Geometric Regressor into precise 3D gaze position for HCI control.

Proposed System

!Image

  • Camera captures real-time eye video feed.
  • DL Gaze Detector estimates gaze direction.
  • Geometric Regressor refines 3D gaze position.
  • Enables precise HCI control interface.

Source: EyeScribe: Vision-Based Assistive HCI

Speaker Notes
High-level flowchart showing Camera → DL Gaze Detector → Geometric Regressor → HCI Control with icons. Illustrates the core pipeline for gaze-based interaction.
Slide 13 - Proposed System
Slide 14 of 17

Slide 14 - Methodology

The slide presents a four-step workflow for gaze tracking methodology. It covers eye/pupil detection with YOLOv5, 3D gaze vector estimation using CNN, 2D point-of-regard prediction via polynomial regression, and output smoothing with a Kalman filter.

Methodology

{ "headers": [ "Step", "Description", "Technology" ], "rows": [ [ "1. Eye/Pupil Detection", "Detect eyes and pupils from real-time video input", "YOLOv5" ], [ "2. Gaze Vector Estimation", "Compute 3D gaze direction vector from eye images", "CNN" ], [ "3. Point-of-Regard Prediction", "Map gaze vector to 2D screen coordinates", "Polynomial Regression" ], [ "4. Smoothing & Output", "Apply temporal filtering for stable gaze tracking", "Kalman Filter" ] ] }

Source: EyeScribe: A Vision-Based Assistive Human-Computer Interface Using Deep Learning and Geometric Regression

Speaker Notes
Walk through the step-by-step workflow: Start with robust eye detection using YOLOv5, estimate gaze vectors with CNN, predict point-of-regard via polynomial regression, and smooth outputs with Kalman filter for precise, real-time HCI.
Slide 14 - Methodology
Slide 15 of 17

Slide 15 - Data Used

The "Data Used" slide table lists three gaze datasets with their sizes and accuracies: MPIIGaze (15,000 images, 94.5%), Eyewriter (custom ~2,500 images, 92.1%), and Combined (~17,500 images, 93.8%). This summarizes the training data performance metrics.

Data Used

{ "headers": [ "Dataset", "Size", "Accuracy (%)" ], "rows": [ [ "MPIIGaze", "15,000 images", "94.5" ], [ "Eyewriter", "Custom (~2,500 images)", "92.1" ], [ "Combined", "~17,500 images", "93.8" ] ] }

Speaker Notes
Datasets include MPIIGaze (15K images) and custom Eyewriter. Used 80/20 train/validation split with augmentations for lighting and pose variations.
Slide 15 - Data Used
Slide 16 of 17

Slide 16 - Features

This slide showcases a grid of five key features for an eye-tracking system, highlighted with icons. They include real-time tracking, head-pose invariance, calibration-free mode, multi-platform support, and accessibility API integration for seamless, inclusive use.

Features

{ "features": [ { "icon": "⚡", "heading": "Real-Time Tracking", "description": "Instant eye movement tracking for seamless, responsive interactions." }, { "icon": "🎭", "heading": "Head-Pose Invariant", "description": "Functions effectively regardless of head position or orientation." }, { "icon": "✅", "heading": "Calibration-Free Mode", "description": "No calibration needed, allowing immediate and effortless setup." }, { "icon": "💻", "heading": "Multi-Platform Support", "description": "Compatible with web and desktop environments for broad accessibility." }, { "icon": "♿", "heading": "Accessibility APIs", "description": "Integrates with OS APIs to enhance usability for all users." } ] }

Slide 16 - Features
Slide 17 of 17

Slide 17 - Tech Stack

The Tech Stack slide highlights three key stats: 30 FPS real-time detection powered by YOLOv5. It also shows 95% classification accuracy via ResNet backbone and optimized deployment speed using ONNX Runtime.

Tech Stack

  • 30 FPS: Real-time Detection
  • YOLOv5 Model

  • 95%: Classification Accuracy
  • ResNet Backbone

  • Optimized: Deployment Speed

ONNX Runtime Source: EyeScribe Project

Speaker Notes
Highlight the technologies used: React.js for frontend, Python/TensorFlow backend, YOLOv5 and ResNet models, ONNX deployment. Emphasize real-time performance with 30 FPS and high accuracy.
Slide 17 - Tech Stack

Discover More Presentations

Explore thousands of AI-generated presentations for inspiration

Browse Presentations
Powered by AI

Create Your Own Presentation

Generate professional presentations in seconds with Karaf's AI. Customize this presentation or start from scratch.

Create New Presentation

Powered by Karaf.ai — AI-Powered Presentation Generator