This is a preview of the print version of your report. Please click "print" to continue or "done" to close this window.

done

or Cancel

Document Viewer
 
Similarity Index
8%
What's this?
Similarity by Source
Internet Sources:
7%
Publications:
6%
Student Papers:
4%
exclude quoted exclude bibliography exclude small matches download print
mode:

Download as pdf Why "Eye Tracking" ? With cities getting smarter and services automated, human-computer interaction is an indispensable aspect of everyday life. As most popular medium of interaction are keyboard and mouse, modern software interfaces are designed to work with hands. Thus, these interfaces are rendered unusable to people with hand disabilities. In this paper review, we discuss intrusive and another non-intrusive realtime interaction technologies, which rely on eye and face movements to control a mouse cursor. Using 'off-the shelf' hardware, open source libraries and a normal processor, real time eye-cursor control is achieved in these papers. Although commercial eye trackers are available in market, they are not very practical for using at home due to either high costs(x000$) or difficult setup procedures. Thus, we need to explore systems which use cheap components easily available in market such as webcams, IR leds, goggles and have open source software allowing any user to calibrate/implement his own system. A commonly used algorithm used by non-intrusive eyetrackers is as follows : A commonly used algorithm used by low cost intrusive eyetrackers is as follows : A head gear with IR Leds and a camera is mounted by the subject In IR light pupils appear darker than rest of face as shown in figure below[1] Using image processing pupils are separated from rest of face Head location relative to surface in front is determined during calibration procedure, when user looks at predefined points projected by a laser on the screen in front Head locations and relative pupil positions are recorded in the calibration session and fed to a training module which approximates a map from eye location to gaze location A webcam is mounted on computer screen and IR Leds are placed besides the webcam IR filter of webcam is stripped making pupil extraction easier Face is detected using Haar Features and eyes are detected in face image again using haar features Eyeballs are detected from eyes by thresholding and blob detection During calibration user is required to look upon points displayed on computer screen and face location(F),Face angles(A), eyeball locations(E) are recorded After obtaining {(F),(A),(E)} either a 3D head model is used to approximate gaze location or neural network is trained on data obtained in previous step to derive a relation sdsg All images shown above are from [1] Images shown above are not copied from anywhere Evaluation of a tracker Many open source eye trackers are available with different specialities. For example ITU gaze tracker [1] is an offline system which packs all processing equipment into a backpack allowing the user to carry system anywhere. This system can be utilized to identify gaze patterns in scenarios such as driving, buying in market or painting. Hence it can be used for researches aiming to classify changes in patterns with learning. For example, an experiment in [4] , where a novice and experienced driver drive a car, came to the conclusion, that as learning progresses in driving, gaze moves towards empty spaces on road, whereas a novice driver looks for cars. In 2009 "ITU gaze tracker", an open source low cost system was released in Spain. This intrusive system consists of a head mounted webcam with inbuilt Infrared emitters. In [2] authors rebuilt the system and test it on two different typing applications(Gaze Talk and Star Gazer). Aim of experiment was to present a system: Which can work in generic situations such as walking or placing display at other location in visible range or in darkness with illumination from IR source only Low-cost , easy to build from off the shelf components Should give reasonably accurate results for atleast one gaze typing system In an experiment on 2 female and 5 male subjects, where each subject typed 60 sentences following results were achieved : Overall study concluded that a major problem which cause loss in accuracy was movement of head and a head pose invariant system would be more robust. Also a speed of 6.56 words per minute can be achieved using ITU's gaze tracker which was comparable to a study on commercial systems where 6.26 WPM was achieved, but not comparable so some others where as much as 15 WPM has been achieved. Then a case study of person in late stages of ALS was conducted in which 'one word' per minute rate was achieved. API to interface trackers with software : Another problem in adaption of gaze tracking technology is absence of standard API's which allow communication between gaze trackers and softwares. Till now most gaze trackers have their own custom built softwares or vice versa. In [3] a standard API is presented which abstracts the communication layer and provides sockets for various control parameters such as Left POG, Right POG, Screen Size , Camera size and handles for calibration and configuration. Although this API does not improve speed of gaze tracking in anyway, but it makes writing applications utilizing open gaze trackers easier thus allowing for faster adoption in development communities. The API operates on a client-server model where Tracker acts as a server and applications as client. This allows many applications to simultaneously utilize the data from tracker's transmission. Conclusion As movement of cursor with head pose and eye positions can be more accurately learnt by a person as compared to a machine, cursor control will improve with learning. But this will still not solve the problem of determining gaze because head-pose variation will cause noise resulting in inaccurate POG's. Hence system will work for interactive purposes but not for research as in [4]. Thus, though gaze tracking can be the future mode of interaction among humans and computers it still lacks features such as accuracy and generality. Begin Match to source 1 in source list: Javier San Agustin. References [1] BABCOCK, J. S., AND PELZ, J. B.End Match 2004. Begin Match to source 1 in source list: Javier San Agustin. Building a lightweight eyetracking headgear. In Proceedings of the 2004 symposium on Eye tracking research & applications, ACM,End Match San Antonio, Texas, Begin Match to source 1 in source list: Javier San Agustin. 109-114. [2]End Match JAVIER SAN AGUSTIN, HENRIK SKOVSGAARD , MARIA BARRET, MARTIN TALL, DAN WITZNER, Begin Match to source 3 in source list: http://martintall.com/index.php?action=publicationsEvaluation of a Low-Cost Open-Source Gaze Tracker ,Proceedings of the 2010 symposium on Eye tracking research & applications,End Match ACM , Begin Match to source 3 in source list: http://martintall.com/index.php?action=publicationsAustinEnd Match TX [3] CRAIG HENNESSEY, ANDREW T. DUCHOWSKIY, Expanding the Adoption of Begin Match to source 4 in source list: Javier San Agustin. Eye-gaze inEnd Match Everyday Applications, Begin Match to source 4 in source list: Javier San Agustin. Proceedings of theEnd Match 2010 Begin Match to source 4 in source list: Javier San Agustin. symposium on Eye tracking research & applications,End Match ACM , Austin TX [4] Begin Match to source 2 in source list: http://www.mood.ws/visual_perception/encyclopedia.htmCohen, A. S. (1983). Informationsaufnahme beim Befahren von Kurven, Psychologie für die Praxis 2/83, Bulletin der Schweizerischen Stiftung für Angewandte PsychologieEnd Match