SlideShare una empresa de Scribd logo
1 de 21
Eye-based head gestures

Diako Mardanbegi
Dan Witzner Hansen
Thomas Pederson




                     IT University of Copenhagen
This paper is about…




                  IT University of Copenhagen
Gaze
   pointing




    Click               Dwell-time
               Blink                  Fixation+Saccade
                         (Fixation)



 DbClick
                        Two steps     Fixation+Saccade
              DbBlink   Dwell-time



  More
commands                               Gaze gestures
Gaze gestures
                                Off-screen targets              Eye-writer
 Mostly used for eye typing       [Isokoski 2000]   !     [Wobbrock, et.al 2008]

 Complex patterns are needed

 Unnatural

 Cognitive load

 Loosing focus on object




                                                         IT University of Copenhagen
Gaze
   pointing




    Click              Dwell-time    Fixation+Saccade
               Wink
                        (Fixation)      (e.g., context
                                          switching)

                                                         Head gestures
 DbClick
                       Two steps     Fixation+Saccade     measured by
              DbWink   Dwell-time                         an eye tracker



  More
commands                             Gaze gestures
Video-based
     head gesture recognition




                           [Nonaka 2003]




[Kapoor and Picard 2002]



                                    IT University of Copenhagen
IT University of Copenhagen
Yaw
                                   Linear eye movements




                           Pitch
      Roll




Rotational eye movements

                                   Pt = [L, R]


                                                  IT University of Copenhagen
Basic movements
     P = [L, R]
     P = [(x, y),(Dx, Dy),(r1, r2,.., r8)]


          P          Basic movement
                        classifier
                                                 Hi




                                      IT University of Copenhagen
GESTURES
Continuous gestures      Discrete gestures




                             Repetitive gestures




                              Sweep gesture
Application Examples




                  IT University of Copenhagen
Classification of gestures
 Character:

 Discrete gestures: repeatable and recognizable sequence of
   characters, Cij=Ci Cj




                    Cit                  Gesture
                                                     Gesture
                                        classifier
                          Application
                             state




                                                               IT University of Copenhagen
Fixation on objects + Head
                        movements
i.     fixed-head eye movements
ii.    fixed-gaze eye movements

      Remote eye trackers (Only eye image):
          Using the reflection of a fixed light source (glint)
      Head-mounted eye trackers (eye image + scene image):
          Using the information obtained from the scene image




                                                                  IT University of Copenhagen
Eye tracker
Method implemented on a head-mounted eye tracker:

Accuracy of about 1.5°

Eye/scene images resolution: 640x480

25 frames per second in real time

Pupil detection: feature-based method

Gaze estimation: homography mapping

Detecting the display corners in the scene image




                                                    IT University of Copenhagen
Testing the classifier
14 predefined gestures have been tested
                                           `

8 participants (6 male and 2 female, mean=35.6, SD=9.7)

Task: Looking at a marker on the screen and then asking the user to
do the gestures




Each gesture has been shown 2 times for each participant

                                                      IT University of Copenhagen
4 participants were not              All participants
                                able to perform                     were able



                        4

         False trials   3

                        2

                        1

                        0                                                           `

                                                      Gestures


False trials because:
   Unable to perform predefined gestures
   Simplicity of the classifier
   Unable to fixate on the marker during gesture
                                                                                        IT University of Copenhagen
Experimental applications
           iRecipe                      iiPhone
Read and follow recipes by   Controlling the iPhone emulator
looking and head gestures    by head gestures
 Investigating the participants experience in terms of physical effort
   and the level of difficulty
                                  difficulty   physical effort

              High            5




                              4
                     Answer




                              3




                              2



              Low             1




                                                                 IT University of Copenhagen
Summary
 Detecting head movements using eye and gaze
  information using both pupil and iris pattern.

 Can keep gaze at the object while interacting
 Found reliable gestures that were comfortable and
  easy to do

 Showed two examples of potential use




                                                   IT University of Copenhagen
Future work

 Improving the accuracy of the classifier by learning user
  specific gestures

 Apply method to control everyday real world objects
 Use the method in a remote eye tracker




                                               IT University of Copenhagen
?

Más contenido relacionado

La actualidad más candente

Evolution of Workflow Provenance Information in the Presence of Custom Infere...
Evolution of Workflow Provenance Information in the Presence of Custom Infere...Evolution of Workflow Provenance Information in the Presence of Custom Infere...
Evolution of Workflow Provenance Information in the Presence of Custom Infere...PlanetData Network of Excellence
 
Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream)
Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream) Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream)
Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream) IT Arena
 
Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...
Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...
Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...IDES Editor
 
Computer Vision For Computer Music
Computer Vision For Computer MusicComputer Vision For Computer Music
Computer Vision For Computer MusicJorge Cardoso
 
MULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNING
MULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNINGMULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNING
MULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNINGcsandit
 
Computer vision techniques for interactive art
Computer vision techniques for interactive artComputer vision techniques for interactive art
Computer vision techniques for interactive artJorge Cardoso
 

La actualidad más candente (7)

Evolution of Workflow Provenance Information in the Presence of Custom Infere...
Evolution of Workflow Provenance Information in the Presence of Custom Infere...Evolution of Workflow Provenance Information in the Presence of Custom Infere...
Evolution of Workflow Provenance Information in the Presence of Custom Infere...
 
Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream)
Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream) Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream)
Face Recognition Based on Deep Learning (Yurii Pashchenko Technology Stream)
 
Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...
Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...
Hardware Implementation of Genetic Algorithm Based Digital Colour Image Water...
 
Computer Vision For Computer Music
Computer Vision For Computer MusicComputer Vision For Computer Music
Computer Vision For Computer Music
 
MULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNING
MULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNINGMULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNING
MULTIMODAL BIOMETRICS RECOGNITION FROM FACIAL VIDEO VIA DEEP LEARNING
 
Computer vision techniques for interactive art
Computer vision techniques for interactive artComputer vision techniques for interactive art
Computer vision techniques for interactive art
 
671 679
671 679671 679
671 679
 

Similar a Eye-based head gestures

DeepVO - Towards Visual Odometry with Deep Learning
DeepVO - Towards Visual Odometry with Deep LearningDeepVO - Towards Visual Odometry with Deep Learning
DeepVO - Towards Visual Odometry with Deep LearningJacky Liu
 
Towards Interaction Models Derived From Eye-tracking Data .
Towards Interaction Models Derived From Eye-tracking Data    .Towards Interaction Models Derived From Eye-tracking Data    .
Towards Interaction Models Derived From Eye-tracking Data . jacekg
 
FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去
FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去
FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去KakeruYamasaki
 
ITP Camp Kinect class
ITP Camp Kinect classITP Camp Kinect class
ITP Camp Kinect classatduskgreg
 
Robotic models of active perception
Robotic models of active perceptionRobotic models of active perception
Robotic models of active perceptionDimitri Ognibene
 
Effect of Self-animated Avatars in Virtual Environments
Effect of Self-animated Avatars in Virtual EnvironmentsEffect of Self-animated Avatars in Virtual Environments
Effect of Self-animated Avatars in Virtual Environmentsmukundraj2
 
Autonomous Learning of Robust Visual Object Detection & Identification on a H...
Autonomous Learning of Robust Visual Object Detection & Identification on a H...Autonomous Learning of Robust Visual Object Detection & Identification on a H...
Autonomous Learning of Robust Visual Object Detection & Identification on a H...Juxi Leitner
 
Towards Neuro–Information Science
Towards Neuro–Information ScienceTowards Neuro–Information Science
Towards Neuro–Information Sciencejacekg
 
426 Lecture 9: Research Directions in AR
426 Lecture 9: Research Directions in AR426 Lecture 9: Research Directions in AR
426 Lecture 9: Research Directions in ARMark Billinghurst
 
Biology microscopes-quiz
Biology microscopes-quizBiology microscopes-quiz
Biology microscopes-quizKrupol Phato
 
Human Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon Transform
Human Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon TransformHuman Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon Transform
Human Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon TransformFadwa Fouad
 
CG OpenGL 3D viewing-course 7
CG OpenGL 3D viewing-course 7CG OpenGL 3D viewing-course 7
CG OpenGL 3D viewing-course 7fungfung Chen
 
Action Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene GraphsAction Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene GraphsSangmin Woo
 

Similar a Eye-based head gestures (14)

DeepVO - Towards Visual Odometry with Deep Learning
DeepVO - Towards Visual Odometry with Deep LearningDeepVO - Towards Visual Odometry with Deep Learning
DeepVO - Towards Visual Odometry with Deep Learning
 
Towards Interaction Models Derived From Eye-tracking Data .
Towards Interaction Models Derived From Eye-tracking Data    .Towards Interaction Models Derived From Eye-tracking Data    .
Towards Interaction Models Derived From Eye-tracking Data .
 
FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去
FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去
FingerVisionを用いた触覚センシングと物体検出を同時に実現するための画像修復とノイズ除去
 
ITP Camp Kinect class
ITP Camp Kinect classITP Camp Kinect class
ITP Camp Kinect class
 
Robotic models of active perception
Robotic models of active perceptionRobotic models of active perception
Robotic models of active perception
 
Effect of Self-animated Avatars in Virtual Environments
Effect of Self-animated Avatars in Virtual EnvironmentsEffect of Self-animated Avatars in Virtual Environments
Effect of Self-animated Avatars in Virtual Environments
 
Autonomous Learning of Robust Visual Object Detection & Identification on a H...
Autonomous Learning of Robust Visual Object Detection & Identification on a H...Autonomous Learning of Robust Visual Object Detection & Identification on a H...
Autonomous Learning of Robust Visual Object Detection & Identification on a H...
 
Towards Neuro–Information Science
Towards Neuro–Information ScienceTowards Neuro–Information Science
Towards Neuro–Information Science
 
426 Lecture 9: Research Directions in AR
426 Lecture 9: Research Directions in AR426 Lecture 9: Research Directions in AR
426 Lecture 9: Research Directions in AR
 
Biology microscopes-quiz
Biology microscopes-quizBiology microscopes-quiz
Biology microscopes-quiz
 
Human Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon Transform
Human Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon TransformHuman Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon Transform
Human Action Recognition in Videos Employing 2DPCA on 2DHOOF and Radon Transform
 
Ooad Uml
Ooad UmlOoad Uml
Ooad Uml
 
CG OpenGL 3D viewing-course 7
CG OpenGL 3D viewing-course 7CG OpenGL 3D viewing-course 7
CG OpenGL 3D viewing-course 7
 
Action Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene GraphsAction Genome: Action As Composition of Spatio Temporal Scene Graphs
Action Genome: Action As Composition of Spatio Temporal Scene Graphs
 

Último

[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdfhans926745
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Paola De la Torre
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking MenDelhi Call girls
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...HostedbyConfluent
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsMaria Levchenko
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Patryk Bandurski
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAndikSusilo4
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 

Último (20)

[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf[2024]Digital Global Overview Report 2024 Meltwater.pdf
[2024]Digital Global Overview Report 2024 Meltwater.pdf
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men08448380779 Call Girls In Greater Kailash - I Women Seeking Men
08448380779 Call Girls In Greater Kailash - I Women Seeking Men
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
Transforming Data Streams with Kafka Connect: An Introduction to Single Messa...
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
Handwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed textsHandwritten Text Recognition for manuscripts and early printed texts
Handwritten Text Recognition for manuscripts and early printed texts
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
Integration and Automation in Practice: CI/CD in Mule Integration and Automat...
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & Application
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 

Eye-based head gestures

  • 1. Eye-based head gestures Diako Mardanbegi Dan Witzner Hansen Thomas Pederson IT University of Copenhagen
  • 2. This paper is about… IT University of Copenhagen
  • 3. Gaze pointing Click Dwell-time Blink Fixation+Saccade (Fixation) DbClick Two steps Fixation+Saccade DbBlink Dwell-time More commands Gaze gestures
  • 4. Gaze gestures Off-screen targets Eye-writer  Mostly used for eye typing [Isokoski 2000] ! [Wobbrock, et.al 2008]  Complex patterns are needed  Unnatural  Cognitive load  Loosing focus on object IT University of Copenhagen
  • 5. Gaze pointing Click Dwell-time Fixation+Saccade Wink (Fixation) (e.g., context switching) Head gestures DbClick Two steps Fixation+Saccade measured by DbWink Dwell-time an eye tracker More commands Gaze gestures
  • 6. Video-based head gesture recognition [Nonaka 2003] [Kapoor and Picard 2002] IT University of Copenhagen
  • 7. IT University of Copenhagen
  • 8. Yaw Linear eye movements Pitch Roll Rotational eye movements Pt = [L, R] IT University of Copenhagen
  • 9. Basic movements P = [L, R] P = [(x, y),(Dx, Dy),(r1, r2,.., r8)] P Basic movement classifier Hi IT University of Copenhagen
  • 10. GESTURES Continuous gestures Discrete gestures Repetitive gestures Sweep gesture
  • 11. Application Examples IT University of Copenhagen
  • 12. Classification of gestures  Character:  Discrete gestures: repeatable and recognizable sequence of characters, Cij=Ci Cj Cit Gesture Gesture classifier Application state IT University of Copenhagen
  • 13. Fixation on objects + Head movements i. fixed-head eye movements ii. fixed-gaze eye movements Remote eye trackers (Only eye image):  Using the reflection of a fixed light source (glint) Head-mounted eye trackers (eye image + scene image):  Using the information obtained from the scene image IT University of Copenhagen
  • 14. Eye tracker Method implemented on a head-mounted eye tracker: Accuracy of about 1.5° Eye/scene images resolution: 640x480 25 frames per second in real time Pupil detection: feature-based method Gaze estimation: homography mapping Detecting the display corners in the scene image IT University of Copenhagen
  • 15. Testing the classifier 14 predefined gestures have been tested ` 8 participants (6 male and 2 female, mean=35.6, SD=9.7) Task: Looking at a marker on the screen and then asking the user to do the gestures Each gesture has been shown 2 times for each participant IT University of Copenhagen
  • 16. 4 participants were not All participants able to perform were able 4 False trials 3 2 1 0 ` Gestures False trials because:  Unable to perform predefined gestures  Simplicity of the classifier  Unable to fixate on the marker during gesture IT University of Copenhagen
  • 17. Experimental applications iRecipe iiPhone Read and follow recipes by Controlling the iPhone emulator looking and head gestures by head gestures
  • 18.  Investigating the participants experience in terms of physical effort and the level of difficulty difficulty physical effort High 5 4 Answer 3 2 Low 1 IT University of Copenhagen
  • 19. Summary  Detecting head movements using eye and gaze information using both pupil and iris pattern.  Can keep gaze at the object while interacting  Found reliable gestures that were comfortable and easy to do  Showed two examples of potential use IT University of Copenhagen
  • 20. Future work  Improving the accuracy of the classifier by learning user specific gestures  Apply method to control everyday real world objects  Use the method in a remote eye tracker IT University of Copenhagen
  • 21. ?

Notas del editor

  1. AVALESH AROOM SHOO KON
  2. In the keynote on wendsday Andrew talked about accelerometers in the smartphones . I’m gona talk about the natural accelerometer that we have in our body call vestibular system that sencethe head movement. The basic idea of this paper is that when you look at an object and move your head your eyes move in the oposit direction because of the vistibulo ocular reflexThis reflexive eye-movements stabilize the eyes in space when the head moves, providing a stable image on the retina..
  3. gaze based interaction basicly started by gaze pointing. However The point of regard does not provide sufficient information for interactingwith interfaces, and this is known as the midas touch problemIntentional blinking and dwell time are typical ways of sending 1 bit information (that can be used for clicking). Two steps dwell time and Blinking twice are also used when we need some more commands (such as dbclickRclick).However, interaction with blinking especially repetitive blinking for long-term use is not very cofrtable. Beside that in some applications we need more command and these methods do not provide more commands. Another method is to use saccades+fixations.Using the intentional eye movements for interaction is known as gaze gestures which can be defined as definable patterns of eye movements performed within a limited time interval
  4. Gaze gestures are mostly used for eye typing.As far as we know it strated by isokosi by the off-screen targets. Where ehe eye gaze has to visit the off-screen targets in a certain order to select characters. Eye writer…In terms of interaction gaze gestures are facing several limitations, for example:Simple gaze gestures can not be distinguished from natural eye patterns. Therefore Complex gaze gestures consist of several simple gaze gestures are needed for robust results. Using the perceptual channel such as vision for motor control may be considered unnatural Besides, the user needs to memorize combinations of gaze gestures,and therefore takes the focus away from the actual interaction task, and increases the cognitive load.The other limitation is that when performing the gaze gestures, the point of regard leaves the object while interacting. So it is not a good method for selecting the icons or interaction with real objectsPEOPLE CAN EASILY GET UNCOMFORTABLE BY DOING GESTURE WHEN I WANT to control an object, I DON’T LIKE TO LOOK SOMEWHERE ELSE
  5. In contrast we sugest to use the head gestures.We offtenuse the head gesturesin our daily life for communication. WE DO NOT USE GAZE GESTURES FOR COMMUNICATIONmost gaze based interactive applications are intended for desabled people who can only move their eyes. Why shouldn’t eye trackers be used for the general population. For example People can move their heads and this can be used for interaction.
  6. Head gesture recognition is well known in computer vision. The general problem are not able to separate the head gestures from the natural head movements There have been attepmts to use the eye information….Only shakes and nodsIn another work they have used an additional sensor (motion tracker) instead of the natural accelerometer. They have used motion tracker together with eye tracker. Eye tracker for … and motion tracker for ….
  7. In contrast, the presented method in this paper, allows for identifying a wide range of head movements even the smallmovemnts accurately and in real time, by only using an eye tracker. What we do is only use the information observed by the cameras of the eye tracker. Here I am showing the example of using the head mounted, but the method applies TO remote eye trackers as well.
  8. There are Three canals in our inner ear that sences the roll, picth and yaw of the head and reflect these movements to the eyes. (for stabilizing the image on the retina)Yaw and pitch create what we call linear movements and roll create what we call rotational..Therefore we can extract a set of two types of feature vectors from the eye image. One part of the the feature vectors are for detecting the rotational eye movements (R) and the other one for linear (L) (pi at time t)
  9. The first part of the feature vector is the pupil center and velosity of the center. the second part of the feature vector which is for detecting the eye rotations. IN OUR IMPLEMENTATION WE HAVE USEd THE MEAN OF THE OPTIC FLOW WITHIN 8 regions around the pupilNow we have the feature vector. We have used a classifier for detecting the basic head movements between two frames of the video sequence by using the feature vector.In this image we the basic movements of head & corresponding Basic reflexive movements of the eyeWe have defined different head gesture using the basic head movements
  10. We have defined two types of gestures:Continuous gestures are defined as the sequences of basic eye movements along an axis Repetitive gestures are where you move the head twice in a same directionSweep when you move the head in one direction and return in the oposit direction. Discrete gstures can be horizental, vertical or diagonal
  11. You can change the volume by continiusly moving the head up and down
  12. Continiuos…Measuring sequences of eye movement are usually influenced by noise. We define a character, Ci, as a sequence of N small eye movements between each two frames where the majority of movements are the same Simple discretegestures,Gij=CiCj are 2 character wordsThere are in total 64 but some of them are not phisicly possible so we only use 14
  13. There are Two types eye movements:So we need a method for detecting whether the gaze is fixed on an object ORFor RT we already have the reflection of a fixed light sourceFor the HMET we need to analize the scene image. In this paper we have used this idea in slitly more dificult setting by using a head mounted eye tracker. The intended use is to apply this technique for gaze based interaction in highly mobile settings.The main advantage of this method with compare to the gaze gestures is that the gaze does not leave the object during the interationWe assume that there is a computer display….
  14. In our own eye tracker
  15. WE HAVE MADE AN EXPERIMENT TO INVESTIGATE THE ACCURACY O FTHE CLASSIFIER. We defined 14 …. As shown here (describe up down diagonal…)We use 8 participants.The method and gestures were introduced to participants and they had the chance of practicing the gestures for 10 minutes before the experiments. 14 simple gestures were shown on the screen by a simple figure, two times one by one and randomly. The shown gesture remains on the screen until the user performs the same gesture or pressing a key in the case when the user was not able to perform that gesture. Each time that a participant performs a gesture but it is not recognized correctly, it will considered as a false trial.
  16. Horizental:..Vertical: average number of false trials of all participants for each gesture.Our general observations were4 participants were not able to perform the diagonal gestures meaning that they were unnaturalRepetitive down gesture is inconvenient
  17. We made two examples to show the potential use of this method for interaction. Four different sweep gestures including Up, Down, Left and Right together with the continuous vertical head movements were used for controlling two applications. Each gesture is interpreted differently based on the gazed object. iRecipe, is intended for a hands-free interaction with a recipe when cooking and when the hands are occupied.iiPhonewhich is an iPhone emulator running on the screen that can be controlled by head gestures to show the potential of thismethod for the mobile devices. WE HAVE THREE REGIONS. A MUSIC IS PLYING DURING COOKING. TO CONTROL THE VOLUME THE USER FIXATES ON THE ….IN THE IIPHONE SWEEPING BETWEEN THE PAGES..We did not meature the false trials but what we observed was that all participants did the tasks susefuly.It is easier to keep the gaze fixed on a meaningful object Participants more liked the volume control by continuous gesture because of the real time sound feedback
  18. After the test, the participants were given a questionnaire consists of questions with the range of the answers from 1 to 5 to investigate the participants experience in terms of physical effort and the level of difficulty of THREE DIFERENT TYPES OF GESTURESthis is the avarage score of different questions which all were about the level of difficulty…..
  19. Thank you very much