Finding the eye(s) in an image is a critical rst step in a remote gaze-tracking system with a working volume large enough to encompass head movements occurring during normal user behavior. We briey review an optical method which exploits the retroreective properties of the eye, and present a novel method for combining difference images to reject motion artifacts. Best performance is obtained when a curvatur operator is used to enhance punctate features, and search is restricted to a neighborhood about the last known location. Optimal setting of the size of this neighborhood is aided by a statistical model of naturally-occurring head movements; we present head-movement statistics mined from a corpus of around 800 hours of video, collected in a team-performance experiment.
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeKalle
Recent advances in high magnification retinal imaging have allowed for visualization of individual retinal photoreceptors, but these systems also suffer from distortions due to fixational eye motion. Algorithms developed to remove these distortions have the added benefit of providing arc second level resolution of the eye movements that produce them. The system also allows for visualization of targets on the retina, allowing for absolute retinal position measures to the level of individual cones. This paper will describe the process used to remove the eye movement artifacts and present analysis of their spectral characteristics. We find a roughly 1/f amplitude spectrum similar to that reported by Findlay (1971) with no evidence for a distinct
tremor component.
Hansen Homography Normalization For Robust Gaze Estimation In Uncalibrated Se...Kalle
Homography normalization is presented as a novel gaze estimation method for uncalibrated setups. The method applies when head movements are present but without any requirements to camera calibration or geometric calibration. The method is geometrically and empirically demonstrated to be robust to head pose changes and despite being less constrained than cross-ratio methods, it consistently
performs favorably by several degrees on both simulated data and data from physical setups. The physical setups include the use of off-the-shelf web cameras with infrared light (night vision) and standard cameras with and without infrared light. The benefits of homography normalization and uncalibrated setups in general are also demonstrated through obtaining gaze estimates (in the visible spectrum) using only the screen reflections on the cornea.
As Digital Still Cameras (DSC) become smaller, cheaper and higher in resolution, photographs are increasingly prone to blurring from shaky hands. Optical image stabilization (OIS) is an effective solution that addresses the quality of images, and is an idea that has been around for at least 30 years. It has only recently made its way into the low-cost consumer camera market, and will soon be migrating to the higher end camera phones. This paper provides an overview of common design practices and considerations for optical image stabilization and how silicon-based MEMS dual-axis gyroscopes with their size, cost and performance advantages are enabling this vital function for image capturing devices
Event recognition image & video segmentationeSAT Journals
Abstract This paper gives a clear look at the segmentation process at the basic level. Segmentation is done at multiple levels so that we will get different results. Segmentation of relative motion descriptors gives a clear picture about the segmentation done for a given input video. Relative motion computation and histograms incrementation are used to evaluate this approach. Also here we will give complete information about the related research which is done about how segmentation can be done for the both images and videos. Keywords: Image Segmentation, Video Segmentation.
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeKalle
Recent advances in high magnification retinal imaging have allowed for visualization of individual retinal photoreceptors, but these systems also suffer from distortions due to fixational eye motion. Algorithms developed to remove these distortions have the added benefit of providing arc second level resolution of the eye movements that produce them. The system also allows for visualization of targets on the retina, allowing for absolute retinal position measures to the level of individual cones. This paper will describe the process used to remove the eye movement artifacts and present analysis of their spectral characteristics. We find a roughly 1/f amplitude spectrum similar to that reported by Findlay (1971) with no evidence for a distinct
tremor component.
Hansen Homography Normalization For Robust Gaze Estimation In Uncalibrated Se...Kalle
Homography normalization is presented as a novel gaze estimation method for uncalibrated setups. The method applies when head movements are present but without any requirements to camera calibration or geometric calibration. The method is geometrically and empirically demonstrated to be robust to head pose changes and despite being less constrained than cross-ratio methods, it consistently
performs favorably by several degrees on both simulated data and data from physical setups. The physical setups include the use of off-the-shelf web cameras with infrared light (night vision) and standard cameras with and without infrared light. The benefits of homography normalization and uncalibrated setups in general are also demonstrated through obtaining gaze estimates (in the visible spectrum) using only the screen reflections on the cornea.
As Digital Still Cameras (DSC) become smaller, cheaper and higher in resolution, photographs are increasingly prone to blurring from shaky hands. Optical image stabilization (OIS) is an effective solution that addresses the quality of images, and is an idea that has been around for at least 30 years. It has only recently made its way into the low-cost consumer camera market, and will soon be migrating to the higher end camera phones. This paper provides an overview of common design practices and considerations for optical image stabilization and how silicon-based MEMS dual-axis gyroscopes with their size, cost and performance advantages are enabling this vital function for image capturing devices
Event recognition image & video segmentationeSAT Journals
Abstract This paper gives a clear look at the segmentation process at the basic level. Segmentation is done at multiple levels so that we will get different results. Segmentation of relative motion descriptors gives a clear picture about the segmentation done for a given input video. Relative motion computation and histograms incrementation are used to evaluate this approach. Also here we will give complete information about the related research which is done about how segmentation can be done for the both images and videos. Keywords: Image Segmentation, Video Segmentation.
Design of Superlens in the visible range using MetamaterialsAhmed Aslam
During last two decades there have been major advances in high resolution imaging using beyond diffraction limit optics where the meta-materials are playing a dominant role. Use of meta-material slabs can act as ideal lens without any aberrations and focus the evanescent waves for super high resolution .
The paper addresses the simulation studies carried out on such meta-material lenses and the theoretical models used to simulate the Super (ideal) lens.
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Kalle
The portability of an eye tracking system encourages us to develop a technique for estimating 3D point-of-regard. Unlike conventional methods, which estimate the position in the 2D image coordinates of the mounted camera, such a technique can represent richer gaze information of the human moving in the larger area. In this paper, we propose a method for estimating the 3D point-of-regard and a visualization technique of gaze trajectories under natural head movements for the head-mounted device. We employ visual SLAM technique to estimate head configuration and extract environmental information. Even in cases where the head moves dynamically, the proposed method could obtain 3D point-of-regard. Additionally, gaze trajectories are appropriately overlaid on the scene camera image.
The introduction of megapixel or HDTV fixed network cameras has meant greater resolution images but not always better image quality. Maintaining image sharpness, particularly in varying outdoor lighting conditions, has proved challenging. It is a problem that has highlighted the limitations of existing lens alternatives and the need for a better solution.P-Iris not only benefits megapixel cameras but all fixed network cameras. The system optimizes the iris opening under all lighting conditions and the result is images with better contrast, clarity, resolution and depth of field. In short, P-Iris means improved image sharpness and increased image usability for network video surveillance operators.
AOS is an industry leader in diffractive optics design & manufacturing. We offer the best lens design, metrology, and extensive fabrication of diffractive optical elements with high precision and diffraction efficiency. Visit Us
Droege Pupil Center Detection In Low Resolution ImagesKalle
In some situations, high quality eye tracking systems are not affordable. This generates the demand for inexpensive systems built upon non-specialized, off the shelf devices. Investigations show that algorithms developed for high resolution systems do not perform satisfactorily on such lowcost and low resolution systems. We investigate
algorithms specifically tailored to such low resolution input devices, based on combination of different strategies. An approach called gradient direction consensus is introduced and compared to image based correlation with adaptive templates as well as other known methods. The results are compared using synthetic input data with known ground truth.
Image Fusion of Video Images and Geo-localization for UAV ApplicationsIDES Editor
We present in this paper a very fine method for
determining the location of a ground based target when viewed
from an Unmanned Aerial Vehicle (UAV). By determining the
pixel coordinates on the video frame and by using a range
finder the target’s geo-location is determined in the North-
East-Down (NED) frame. The contribution of this method is
that the target can be localized to within 9m when view from
an altitude of 2500m and down to 1m from an altitude of 100m.
This method offers a highly versatile tracking and geolocalisation
technique that has very good number of
advantages over the previously suggested methods. Some of
the key factors that differentiate our method from its
predecessors are:
1) Day and night time operation
2) All weather operation
3) Highly accurate positioning of target in terms of
latitude-longitude (GPS) and altitude.
4) Automatic gimbaled operation of the camera once
target is locked
5) Tracking is possible even when the target stops
moving
6) Independent of target (moving or stationary)
7) No terrain database is required
8) Instantaneous target geolocalisation is possible
Purkinje imaging for crystalline lens density measurementPetteriTeikariPhD
Brief introduction for the non-invasive, inexpensive and fast Purkinje image -based method for measuring the spectral transmittance of the human crystalline lens density in vivo.
Alternative download link:
https://www.dropbox.com/s/588y7epy13n34xo/purkinje_imaging.pdf?dl=0
Presentation of Detlev Breyer (MD) at the ASCRS 2017 in Los Angeles. Title: New refractive optical MIOL concepts – Comparison of different optical systems.
Presentation of Detlev Breyer (MD) at ERLS Meeting in Paris in 2017. Title: MIOL Treatment – Quo Vadis? This presentation discusses the treatment of presbyopia with Multifocal Lenses.
Ultra High Focusing Speed up to 12000Hz
Long Term Reliability more than 1 billion operating cycle
Ultra Small Power Consumption less than 1mA
Shock Resistance more than 5000G
Operating Temperature Range: -30 ~ 100°C
Single Camera Based 3D Camera
Volumetrically 25% Smaller than Current Technology Based Module
Real-Time Multi Focusing
Applicable to Volumetric 3D DISPLAY, Compact Auto Focus and 3D camera module
A maskless exposure device for rapid photolithographic prototyping of sensor ...Dhanesh Rajan
A very cost effective maskless exposure device (MED) for the fast lithographic prototyping of various layouts is presented. The device is assembled using a digital light processing projector (DLP), an optical microscope, alignment stages and a web camera. Layouts created on a computer screen can be easily transferred to substrate surfaces without using expensive photomasks and the process can be repeated by introducing new drawings on the screen. Components are tuned for a constant area of exposure and a resolution of around 20 μm is possible at the moment without using any reduction lenses. The MED has been used in patterning the surfaces of silicon, glass, metal etc. successfully. The device can be assembled using commercially available components at a very minimum cost and can be effectively used in fast prototyping applications like in MEMS, microfluidics, patterning of sensor and electrode structures.
The Light-CT scanner allows to get high resolution, 1 µm in 3D images within minutes.
This works direcly on fresh, fixed or frozen tissue.
No staining is required.
This atlazd shows some of the capabilities of the system and the technoologu applied to human skin in vivo and ex vivo (skin models..)
Original Power Point retrieved from http://www.mrsshirley.net/powerpoint/realidades/grammar/
real1grammar/real1grammar.htm. Educational use granted if credit was given to original author.
Design of Superlens in the visible range using MetamaterialsAhmed Aslam
During last two decades there have been major advances in high resolution imaging using beyond diffraction limit optics where the meta-materials are playing a dominant role. Use of meta-material slabs can act as ideal lens without any aberrations and focus the evanescent waves for super high resolution .
The paper addresses the simulation studies carried out on such meta-material lenses and the theoretical models used to simulate the Super (ideal) lens.
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Kalle
The portability of an eye tracking system encourages us to develop a technique for estimating 3D point-of-regard. Unlike conventional methods, which estimate the position in the 2D image coordinates of the mounted camera, such a technique can represent richer gaze information of the human moving in the larger area. In this paper, we propose a method for estimating the 3D point-of-regard and a visualization technique of gaze trajectories under natural head movements for the head-mounted device. We employ visual SLAM technique to estimate head configuration and extract environmental information. Even in cases where the head moves dynamically, the proposed method could obtain 3D point-of-regard. Additionally, gaze trajectories are appropriately overlaid on the scene camera image.
The introduction of megapixel or HDTV fixed network cameras has meant greater resolution images but not always better image quality. Maintaining image sharpness, particularly in varying outdoor lighting conditions, has proved challenging. It is a problem that has highlighted the limitations of existing lens alternatives and the need for a better solution.P-Iris not only benefits megapixel cameras but all fixed network cameras. The system optimizes the iris opening under all lighting conditions and the result is images with better contrast, clarity, resolution and depth of field. In short, P-Iris means improved image sharpness and increased image usability for network video surveillance operators.
AOS is an industry leader in diffractive optics design & manufacturing. We offer the best lens design, metrology, and extensive fabrication of diffractive optical elements with high precision and diffraction efficiency. Visit Us
Droege Pupil Center Detection In Low Resolution ImagesKalle
In some situations, high quality eye tracking systems are not affordable. This generates the demand for inexpensive systems built upon non-specialized, off the shelf devices. Investigations show that algorithms developed for high resolution systems do not perform satisfactorily on such lowcost and low resolution systems. We investigate
algorithms specifically tailored to such low resolution input devices, based on combination of different strategies. An approach called gradient direction consensus is introduced and compared to image based correlation with adaptive templates as well as other known methods. The results are compared using synthetic input data with known ground truth.
Image Fusion of Video Images and Geo-localization for UAV ApplicationsIDES Editor
We present in this paper a very fine method for
determining the location of a ground based target when viewed
from an Unmanned Aerial Vehicle (UAV). By determining the
pixel coordinates on the video frame and by using a range
finder the target’s geo-location is determined in the North-
East-Down (NED) frame. The contribution of this method is
that the target can be localized to within 9m when view from
an altitude of 2500m and down to 1m from an altitude of 100m.
This method offers a highly versatile tracking and geolocalisation
technique that has very good number of
advantages over the previously suggested methods. Some of
the key factors that differentiate our method from its
predecessors are:
1) Day and night time operation
2) All weather operation
3) Highly accurate positioning of target in terms of
latitude-longitude (GPS) and altitude.
4) Automatic gimbaled operation of the camera once
target is locked
5) Tracking is possible even when the target stops
moving
6) Independent of target (moving or stationary)
7) No terrain database is required
8) Instantaneous target geolocalisation is possible
Purkinje imaging for crystalline lens density measurementPetteriTeikariPhD
Brief introduction for the non-invasive, inexpensive and fast Purkinje image -based method for measuring the spectral transmittance of the human crystalline lens density in vivo.
Alternative download link:
https://www.dropbox.com/s/588y7epy13n34xo/purkinje_imaging.pdf?dl=0
Presentation of Detlev Breyer (MD) at the ASCRS 2017 in Los Angeles. Title: New refractive optical MIOL concepts – Comparison of different optical systems.
Presentation of Detlev Breyer (MD) at ERLS Meeting in Paris in 2017. Title: MIOL Treatment – Quo Vadis? This presentation discusses the treatment of presbyopia with Multifocal Lenses.
Ultra High Focusing Speed up to 12000Hz
Long Term Reliability more than 1 billion operating cycle
Ultra Small Power Consumption less than 1mA
Shock Resistance more than 5000G
Operating Temperature Range: -30 ~ 100°C
Single Camera Based 3D Camera
Volumetrically 25% Smaller than Current Technology Based Module
Real-Time Multi Focusing
Applicable to Volumetric 3D DISPLAY, Compact Auto Focus and 3D camera module
A maskless exposure device for rapid photolithographic prototyping of sensor ...Dhanesh Rajan
A very cost effective maskless exposure device (MED) for the fast lithographic prototyping of various layouts is presented. The device is assembled using a digital light processing projector (DLP), an optical microscope, alignment stages and a web camera. Layouts created on a computer screen can be easily transferred to substrate surfaces without using expensive photomasks and the process can be repeated by introducing new drawings on the screen. Components are tuned for a constant area of exposure and a resolution of around 20 μm is possible at the moment without using any reduction lenses. The MED has been used in patterning the surfaces of silicon, glass, metal etc. successfully. The device can be assembled using commercially available components at a very minimum cost and can be effectively used in fast prototyping applications like in MEMS, microfluidics, patterning of sensor and electrode structures.
The Light-CT scanner allows to get high resolution, 1 µm in 3D images within minutes.
This works direcly on fresh, fixed or frozen tissue.
No staining is required.
This atlazd shows some of the capabilities of the system and the technoologu applied to human skin in vivo and ex vivo (skin models..)
Original Power Point retrieved from http://www.mrsshirley.net/powerpoint/realidades/grammar/
real1grammar/real1grammar.htm. Educational use granted if credit was given to original author.
Original Power Point retrieved from http://www.mrsshirley.net/powerpoint/realidades/grammar/
real1grammar/real1grammar.htm. Educational use granted if credit given to author.
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural TasksKalle
Analysis of recordings made by a wearable eye tracker is complicated by video stream synchronization, pupil coordinate mapping, eye movement analysis, and tracking of dynamic Areas Of Interest (AOIs) within the scene. In this paper a semi-automatic system is developed to help automate these processes. Synchronization is accomplished
via side by side video playback control. A deformable eye template and calibration dot marker allow reliable initialization via simple drag and drop as well as a user-friendly way to correct the algorithm when it fails. Specifically, drift may be corrected by nudging the detected pupil center to the appropriate coordinates. In a case study, the impact of surrogate nature views on physiological health and perceived well-being is examined via analysis of gaze over images of nature. A match-moving methodology was developed to track AOIs for this particular application but is applicable toward similar future studies.
Eye Gaze Tracking With a Web Camera in a Desktop Environment1crore projects
IEEE PROJECTS 2015
1 crore projects is a leading Guide for ieee Projects and real time projects Works Provider.
It has been provided Lot of Guidance for Thousands of Students & made them more beneficial in all Technology Training.
Dot Net
DOTNET Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
Java Project Domain list 2015
1. IEEE based on datamining and knowledge engineering
2. IEEE based on mobile computing
3. IEEE based on networking
4. IEEE based on Image processing
5. IEEE based on Multimedia
6. IEEE based on Network security
7. IEEE based on parallel and distributed systems
ECE IEEE Projects 2015
1. Matlab project
2. Ns2 project
3. Embedded project
4. Robotics project
Eligibility
Final Year students of
1. BSc (C.S)
2. BCA/B.E(C.S)
3. B.Tech IT
4. BE (C.S)
5. MSc (C.S)
6. MSc (IT)
7. MCA
8. MS (IT)
9. ME(ALL)
10. BE(ECE)(EEE)(E&I)
TECHNOLOGY USED AND FOR TRAINING IN
1. DOT NET
2. C sharp
3. ASP
4. VB
5. SQL SERVER
6. JAVA
7. J2EE
8. STRINGS
9. ORACLE
10. VB dotNET
11. EMBEDDED
12. MAT LAB
13. LAB VIEW
14. Multi Sim
CONTACT US
1 CRORE PROJECTS
Door No: 214/215,2nd Floor,
No. 172, Raahat Plaza, (Shopping Mall) ,Arcot Road, Vadapalani, Chennai,
Tamin Nadu, INDIA - 600 026
Email id: 1croreprojects@gmail.com
website:1croreprojects.com
Phone : +91 97518 00789 / +91 72999 51536
Highly Efficient Immersive AR Glasses with Gaze-locked Exit Pupil Steering high level architecture and technology explanation given by Boris Greenberg at SPIE 2021
Model User Calibration Free Remote Gaze Estimation SystemKalle
Gaze estimation systems use calibration procedures that require active subject participation to estimate the point-of-gaze accurately. In these procedures, subjects are required to fixate on a specific point or points in space at specific time instances. This
paper describes a gaze estimation system that does not use calibration procedures that require active user participation. The
system estimates the optical axes of both eyes using images from a stereo pair of video cameras without a personal calibration procedure. To estimate the point-of-gaze, which lies along the visual axis, the angles between the optical and visual axes are estimated by a novel automatic procedure that minimizes the distance between the intersections of the visual axes of the left and right eyes with the surface of a display while subjects look naturally at the display (e.g., watching a video clip). Experiments
with four subjects demonstrate that the RMS error of this point-of-gaze estimation system is 1.3º.
Automatic identification of animal using visual and motion saliencyeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Nagamatsu User Calibration Free Gaze Tracking With Estimation Of The Horizont...Kalle
This paper presents a user-calibration-free method for estimating the point of gaze (POG) on a display accurately with estimation of the horizontal angles between the visual and the optical axes of both eyes. By using one pair of cameras and two light sources, the
optical axis of the eye can be estimated. This estimation is carried
out by using a spherical model of the cornea. The point of intersection
of the optical axis of the eye with the display is termed POA.By detecting the POAs of both the eyes, the POG is approximately estimated as the midpoint of the line joining the POAs of both the eyes on the basis of the binocular eye model; therefore, we can estimate
the horizontal angles between the visual and the optical axes of both the eyes without requiring user calibration. We have developed
a prototype system based on this method using a 19 display with two pairs of stereo cameras. We evaluated the system experimentally with 20 subjects who were at a distance of 600 mm from the display. The result shows that the average of the root-meansquare error (RMSE) of measurement of POG in the display screen coordinate system is 16.55 mm (equivalent to less than 1.58◦).
Deconstructing SfM-Net architecture and beyond
"SfM-Net, a geometry-aware neural network for motion estimation in videos that decomposes frame-to-frame pixel motion in terms of scene and object depth, camera motion and 3D object rotations and translations. Given a sequence of frames, SfM-Net predicts depth, segmentation, camera and rigid object motions, converts those into a dense frame-to-frame motion field (optical flow), differentiably warps frames in time to match pixels and back-propagates."
Alternative download:
https://www.dropbox.com/s/aezl7ro8sy2xq7j/sfm_net_v2.pdf?dl=0
Effective Object Detection and Background Subtraction by using M.O.IIJMTST Journal
This paper proposes efficient motion detection and people counting based on background subtraction using dynamic threshold approach with mathematical morphology. Here these different methods are used effectively for object detection and compare these performance based on accurate detection. Here the techniques frame differences, dynamic threshold based detection will be used. After the object foreground detection, the parameters like speed, velocity motion will be determined. For this, most of previous methods depend on the assumption that the background is static over short time periods. In dynamic threshold based object detection, morphological process and filtering also used effectively for unwanted pixel removal from the background. The background frame will be updated by comparing the current frame intensities with reference frame. Along with this dynamic threshold, mathematical morphology also used which has an ability of greatly attenuating color variations generated by background motions while still highlighting moving objects. Finally the simulated results will be shown that used approximate median with mathematical morphology approach is effective rather than prior background subtraction methods in dynamic texture scenes and performance parameters of moving object such sensitivity, speed and velocity will be evaluated by using MOI.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Similar to Mulligan Robust Optical Eye Detection During Head Movement (20)
Blignaut Visual Span And Other Parameters For The Generation Of HeatmapsKalle
Although heat maps are commonly provided by eye-tracking and visualization tools, they have some disadvantages and caution must be taken when using them to draw conclusions on eye tracking results. It is motivated here that visual span is an essential component of visualizations of eye-tracking data and an algorithm is proposed to allow the analyst to set the visual span as a parameter prior to generation of a heat map.
Although the ideas are not novel, the algorithm also indicates how transparency of the heat map can be achieved and how the color gradient can be generated to represent the probability for an object to be observed within the defined visual span. The optional addition of contour lines provides a way to visualize separate intervals in the continuous color map.
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...Kalle
Relevance feedback (RF) mechanisms are widely adopted in Content-Based Image Retrieval (CBIR) systems to improve image retrieval performance. However, there exist some intrinsic problems: (1) the semantic gap between high-level concepts and low-level features and (2) the subjectivity of human perception of visual contents. The primary focus of this paper is to evaluate the possibility of inferring the relevance of images based on eye movement data. In total, 882 images from 101 categories are viewed by 10 subjects to test the usefulness of implicit RF, where the relevance of each image is known beforehand. A set of measures based on fixations are thoroughly evaluated which include fixation duration, fixation count, and the number of revisits. Finally, the paper proposes a decision tree to predict the user’s input during the image searching tasks. The prediction precision of the decision tree is over 87%, which spreads light on a promising integration of natural eye movement into CBIR systems in the future.
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...Kalle
The intuitive user interfaces of PCs and PDAs, such as pen display and touch panel, have become widely used in recent times. In this study, we have developed an eye-tracking pen display based on the stereo bright pupil technique. First, the bright pupil camera was developed by examining the arrangement of cameras and LEDs for pen display. Next, the gaze estimation method was proposed for the stereo bright pupil camera, which enables one point calibration. Then, the prototype of the eyetracking pen display was developed. The accuracy of the system was approximately 0.7° on average, which is sufficient for human interaction support. We also developed an eye-tracking tabletop as an application of the proposed stereo bright pupil technique.
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...Kalle
Individuals with severe multiple disabilities have little or no opportunity to express their own wishes, make choices and move independently. Because of this, the objective of this work has been to develop a prototype for a gaze-driven device to manoeuvre powered wheelchairs or other moving platforms. The prototype has the same capabilities as a normal powered wheelchair, with two exceptions. Firstly, the prototype is controlled by eye movements instead of by a normal joystick. Secondly, the prototype is equipped with a sensor that stops all motion when the machine approaches an obstacle. The prototype has been evaluated in a preliminary clinical test with two users. Both users clearly communicated that they appreciated and had mastered the ability to control a powered wheelchair with their eye movements.
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...Kalle
The visual field is the area of space that can be seen when an observer fixates a given point. Many visual capabilities vary with position in the visual field and many diseases result in changes in the visual field. With current technology, it is possible to build very complex real-time visual field simulations that employ gaze-contingent displays. Nevertheless, there are still no established techniques to evaluate such systems. We have developed a method to evaluate a system’s contingency by employing visual blind spot localization as well as foveal fixation. During the experiment, gaze-contingent and static conditions were compared. There was a strong correlation between predicted results and gaze-contingent trials. This evaluation method can also be used with patient populations and for the evaluation of gaze-contingent display systems, when there is need to evaluate a visual field outside of the foveal region.
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze ControlKalle
Pie menus offer several features which are advantageous especially for gaze control. Although the optimal number of slices per pie
and of depth layers has already been established for manual control, these values may differ in gaze control due to differences in spatial accuracy and congitive processing. Therefore, we investigated the layout limits for hierarchical pie menu in gaze control. Our user study indicates that providing six slices in multiple depth layers guarantees fast and accurate selections. Moreover, we compared two different methods of selecting a slice. Novices performed well with both, but selecting via selection borders produced better performance for experts than the standard dwell time selection.
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...Kalle
Eye typing could provide motor disabled people a reliable method of communication given that the text entry speed of current interfaces can be increased to allow for fluent communication. There are two reasons for the relatively slow text entry: dwell time selection requires waiting a certain time, and single character entry limits the maximum entry speed. We adopted a typing interface based on hierarchical pie menus, pEYEwrite [Urbina and Huckauf 2007] and included bigram text entry with one single pie iteration. Therefore, we introduced three different bigram building strategies.
Moreover, we combined dwell time selection with selection by borders, providing an alternative selection method and extra functionality. In a longitudinal study we compared participants performance during character-by-character text entry with bigram entry and with
text entry with bigrams derived by word prediction. Data showed large advantages of the new entry methods over single character text entry in speed and accuracy. Participants preferred selecting by
borders, which allowed them faster selections than the dwell time method.
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic TrainingKalle
The study of surgeons’ eye movements is an innovative way of assessing skill and situation awareness, in that a comparison of eye movement strategies between expert surgeons and novices may show differences that can be used in training. Our preliminary study compared eye movements of 4 experts and
4 novices performing a simulated gall bladder removal task on a
dummy patient with an audible heartbeat and simulated vital signs displayed on a secondary monitor. We used a head-mounted Locarna PT-Mini eyetracker to record fixation locations during the operation. The results showed that novices concentrated so hard on the surgical
display that they were hardly able to look at the patient’s vital signs, even when heart rate audibly changed during the procedure. In comparison, experts glanced occasionally at the vitals monitor, thus being able to observe the patient condition.
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...Kalle
Gaze visualizations represent an effective way for gaining fast insights into eye tracking data. Current approaches do not adequately support eye tracking studies for three-dimensional (3D) virtual environments. Hence, we propose a set of advanced gaze visualization techniques for supporting gaze behavior analysis in such environments. Similar to commonly used gaze visualizations for twodimensional
stimuli (e.g., images and websites), we contribute advanced 3D scan paths and 3D attentional maps. In addition, we introduce a models of interest timeline depicting viewed models, which can be used for displaying scan paths in a selected time segment. A prototype toolkit is also discussed which combines an implementation of our proposed techniques. Their potential for facilitating eye tracking studies in virtual environments was supported by a user study among eye tracking and visualization experts.
Skovsgaard Small Target Selection With Gaze AloneKalle
Accessing the smallest targets in mainstream interfaces using gaze
alone is difficult, but interface tools that effectively increase the size of selectable objects can help. In this paper, we propose a conceptual framework to organize existing tools and guide the development of new tools. We designed a discrete zoom tool and conducted a proof-of-concept experiment to test the potential of the framework and the tool. Our tool was as fast as and more accurate than the currently available two-step magnification tool. Our framework shows potential to guide the design, development, and testing of zoom tools to facilitate the accessibility of mainstream
interfaces for gaze users.
San Agustin Evaluation Of A Low Cost Open Source Gaze TrackerKalle
This paper presents a low-cost gaze tracking system that is based on a webcam mounted close to the user’s eye. The performance of the gaze tracker was evaluated in an eye-typing task using two different typing applications. Participants could type between 3.56 and 6.78 words per minute, depending on the typing system used. A pilot study to assess the usability of the system was also carried out in the home of a user with severe motor impairments. The
user successfully typed on a wall-projected interface using his eye movements.
Rosengrant Gaze Scribing In Physics Problem SolvingKalle
Eye-tracking has been widely used for research purposes in fields such as linguistics and marketing. However, there are many possibilities of how eye-trackers could be used in other disciplines like physics. A part of physics education research deals with the differences between novices and experts, specifi-cally how each group solves problems. Though there has been a great deal of research about these differences there has been no research that focuses on noticing exactly where experts and no-vices look while solving the problems. Thus, to complement the past research, I have created a new technique called gaze scrib-ing. Subjects wear a head mounted eye-tracker while solving electrical circuit problems on a graphics monitor. I monitor both scan patterns of the subjects and combine that with videotapes of their work while solving the problems. This new technique has yielded new information and elaborated on previous studies.
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual SearchKalle
In certain applications such as radiology and imagery analysis, it is important to minimize errors. In this paper we evaluate a structured inspection method that uses eye tracking information as a feedback mechanism to the image inspector. Our two-phase method starts with a free viewing phase during which gaze data is collected. During the next phase, we either segment the image, mask previously seen areas of the image, or combine the two techniques, and repeat the search. We compare the different methods
proposed for the second search phase by evaluating the inspection method using true positive and false negative rates, and subjective workload. Results show that gaze-blocked configurations reduced the subjective workload, and that gaze-blocking without segmentation showed the largest increase in true positive identifications and the largest decrease in false negative identifications of previously unseen objects.
Prats Interpretation Of Geometric Shapes An Eye Movement StudyKalle
This paper describes a study that seeks to explore the correlation between eye movements and the interpretation of geometric shapes. This study is intended to inform the development of an eye tracking interface for computational tools to support and enhance the natural interaction required in creative design. A common criticism of computational design tools is that they do not enable manipulation of designed shapes according to all perceived features. Instead the manipulations afforded are limited by formal structures of shapes. This research examines the potential for eye movement data to be used to recognise and make available for manipulation the perceived features in shapes. The objective of this study was to analyse eye movement data with the intention of recognising moments in which an interpretation of shape is made. Results suggest that fixation duration and saccade amplitude prove to be consistent indicators of shape interpretation.
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...Kalle
Eye gaze interaction for disabled people is often dealt with by designing ad-hoc interfaces, in which the big size of their elements compensates for both the inaccuracy of eye trackers and the instability of the human eye. Unless solutions for reliable eye cursor control are employed, gaze pointing in ordinary graphical operating environments is a very difficult task. In this paper we present an eye-driven cursor for MS Windows which behaves differently according to the “context”. When the user’s gaze is perceived within the desktop or a folder, the cursor can be discretely shifted from one icon to another. Within an application window or where there are no icons, on the contrary, the cursor can be continuously and precisely moved. Shifts in the four directions (up, down, left, right) occur through dedicated buttons. To increase user awareness of the currently pointed spot on the screen while continuously moving the cursor, a replica of the spot is provided within the active direction button, resulting in improved pointing performance.
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...Kalle
Laboratory eyetrackers, constrained to a fixed display and static (or accurately tracked) observer, facilitate automated analysis of fixation data. Development of wearable eyetrackers has extended environments and tasks that can be studied at the expense of automated analysis. Wearable eyetrackers provide 2D point-of-regard (POR) in scene-camera coordinates, but the researcher is typically interested in some high-level semantic property (e.g., object identity, region, or material) surrounding individual fixation points. The synthesis of POR into fixations and semantic information remains a labor-intensive manual task, limiting the application of wearable eyetracking.
We describe a system that segments POR videos into fixations and allows users to train a database-driven, object-recognition system. A correctly trained library results in a very accurate and semi-automated translation of raw POR data into a sequence of objects, regions or materials.
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...Kalle
The purpose of this study is to explore how the viewers’ previous training is related to their aesthetic viewing in various interactions with the form and the context, in relation to apparel design. Berlyne’s two types of exploratory behavior, diversive and specific, provided a theoretical framework to this study. Twenty female subjects (mean age=21, SD=1.089) participated. Twenty model images, posed by a male and a female model, were shown on an eye-tracker screen for 10 seconds each. The findings of this study verified Berlyne’s concepts of visual exploration. One of the different findings from Berlyne’s theory was that the untrained viewers’ visual attention tended to be more significantly focused on peripheral areas of visual interest, compared to the trained viewers, while there was no significant difference on the central, foremost areas of visual interest between the two groups. The overall aesthetic viewing patterns were also identified.
Palinko Estimating Cognitive Load Using Remote Eye Tracking In A Driving Simu...Kalle
We report on the results of a study in which pairs of subjects were involved in spoken dialogues and one of the subjects also operated a simulated vehicle. We estimated the driver’s cognitive load based on pupil size measurements from a remote eye tracker. We compared the cognitive load estimates based on the physiological pupillometric data and driving performance data. The physiological and performance measures show high correspondence suggesting that remote eye tracking might provide reliable driver cognitive load estimation, especially in simulators. We also introduced a new pupillometric cognitive load measure that shows promise in tracking cognitive load changes on time scales of several seconds.
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...Kalle
To estimate viewer’s contextual understanding, features of their
eye-movements while viewing question statements in response to definition statements, and features of correct and incorrect responses were extracted and compared. Twelve directional features
of eye-movements across a two-dimensional space were created, and these features were compared between correct and incorrect responses. The procedure of estimating the response was developed with Support Vector Machines, using these features. The estimation performance and accuracy were assessed across combinations of features. The number of definition statements, which needed to be memorized to answer the question statements during the experiment, affected the estimation accuracy. These results provide evidence that features of eye-movements during reading statements
can be used as an index of contextual understanding.
Nagamatsu Gaze Estimation Method Based On An Aspherical Model Of The Cornea S...Kalle
A novel gaze estimation method based on a novel aspherical model of the cornea is proposed in this paper. The model is a surface of revolution about the optical axis of the eye. The calculation method is explained on the basis of the model. A prototype system for estimating the point of gaze (POG) has been developed using this method. The proposed method has been found to be more accurate than the gaze estimation method based on a spherical model of the
cornea.
2. Figure 1: A series of images from a wide-field camera equipped with an active illumination system. Bright pupil illumination is provided
during odd fields, while dark pupil illumination is provided during even fields. Within-frame difference images are obtained by subtracting
the two fields from a single frame, while across-frame difference images are obtained by subtracting fields from different frames. These two
types of difference images show different types of motion artifacts, which may be in large part suppressed by taking their product, as shown
in the bottom row. To enhance visibility, all difference images have been inverted, blurred, and normalized.
second (even) field of the previous frame is subtracted from the first
(odd) field of the current frame. The difference images are clipped
from below at zero, in order to suppress regions where the dark wi = max(oi − ei , 0), (1)
pupil image is brighter than the bright pupil image. The difference ai = max(oi − ei−1 , 0), and (2)
images have been inverted and slightly blurred to enhance visibility ci = a i wi , (3)
in the present reproduction.
where the all vector operations are understood to represent the cor-
To get the most up-to-date estimate of the pupil location, we should responding pixel-wise operation on the images.
use the most recently acquired pair of bright- and dark-pupil im- It can be seen in figure 1 that the combined difference image has a
ages. But, as illustrated in figure 1, the temporal ordering of strong signal in the eye regions, and has reduced the strength of the
the bright- and dark-pupil fields alternates, and so the polarity of motion artifacts. The source of the eye signal in this case can be
motion-induced artifacts alternates as well. The half-wave rectifi- quite different than in the case of a stationary head; when the head
cation performed after the differencing suppresses roughly half of and eyes are still, the only difference between the two images will
the motion-induced artifacts; in figure 1 the head is moving to the be the brightness of the pupil region - the glints (often the brightest
left, and so strong motion-induced artifacts are seen at the left edge part of the image) will cancel. When the head and/or eye is in
of the head in the within-frame differences, and on the right edge motion, on the other hand, the glints will not cancel each other; but
of the head in the across-frame differences. Because a particular when our goal is to simply find the eye (as opposed to accurately
head-edge artifact is in general only present in one of the two types segmenting the pupil region), then this is not a problem: because the
of difference image, we are able to eliminate most of them (and glint in the bright pupil image can be expected to be brighter than
reduce the strength of the remainder relative to the eye signals) by anything else with which it might be (mis)aligned, we are likely
pixel-wise multiplication the pair of difference images derived from to get a strong signal in the difference image even when the pupil
a single bright-pupil image. We can express this more formally as signal is degraded.
follows: let oi and ei represent the odd and even fields (respec-
tively) from frame i. We further let wi and ai represent the within- We have performed a preliminary evaluation of the technique using
and across-frame differences, respectively, and let ci represent the a video sequence captured while a subject executed voluntary, side-
combined difference image. Then to-side head movements, ranging from slow to as fast as possible.
130
3. Figure 2: Proportion of frames in which both eyes were correctly Figure 3: Similar to figure 2, for four algorithms in which within-
localized as a function of average head speed, for four algorithms, and across-frame difference images were combined as illustrated
based on: single inter-frame differences (inverted triangles); dif- in figure 1: combined difference images (inverted triangles); com-
ference images with a restricted search region (upright triangles); bined difference images with a restricted search area (upright trian-
Gaussian curvature of difference image (squares); and Gaussian gles); combined curvature images (squares); and combined curva-
curvature with a restricted search region (diamonds). ture images with a restricted search area (diamonds).
The imaging configuration was similar to that shown in figure 1, size of the search window, it is therefore useful to know something
with the head taking up roughly half the width of the frame (about about the expected movements of the subjects. Such data are also
300 pixels), and an effective frame rate of 60 Hz after deinterlacing. useful for determining parameters of a remote gaze tracking sys-
Difference images were computed, and blurred slightly. Our base- tem, such as field-of-view and steering speed. To this end, we an-
line algorithm was to search for the first two local maxima in the alyzed a large corpus of video, provided to us by a research group
blurred difference image. While this simple procedure performs studying team decision-making [Fischer et al. 2007]. In this study,
satisfactorily when the head is still, it breaks down as significant 5-member teams participated in a (simulated) tele-robotic search
head motion is encountered (see figure 2). A couple of simple en- mission, by manning console stations in each of 5 separate exper-
hancements can improve performance somewhat: first we can re- imental chambers. Video footage of each subject’s work session
strict the search area for each eye to a small neighborhood centered was recorded by a video camera mounted above their workstation,
on the previous position (in our example we used a radius of 7 pix- with a horizontal field-of-view of about 40 degrees. Although the
els, larger than the highest head speed). Secondly, we can apply subjects in the videos assume a variety of seated poses while at
a transformation to the difference image which accentuates punc- work, their detected faces typically occupied between roughly one-
tate features while suppressing linear features; this is motivated by quarter to one-third of a frame width (80 - 106 pixels). The corpus
the observation that many of the most severe motion artifacts occur made available to us consisted of 258 separate recordings, compris-
around the edges of the face and are elongated. The transformation ing a total of 793 hours (42.8 million images). The recordings had
we chose was Gaussian curvature [Zetzsche and Barth 1990; Barth already been subjected to some preprocessing; as received by us,
et al. 1998], a combination of second derivatives which is zero for each movie had a resolution of 320 x 240, at a temporal sampling
linear features but responds strongly to spots and corners. We can rate of 15 frames per second.
see in figure 2 that, while application of the curvature operator does
produce a performance benefit, the greatest benefit comes from as- To detect the location of the face in each frame, we used a boosted
suming the eyes are not far from their previous locations. This too cascade of classifiers [Viola and Jones 2001]. We used the imple-
breaks down, however, at moderate head speeds. mentation provided in the OpenCV library [Lienhart and Maydt
2002]. Exactly 1 face was found in approximately 50.9% of the
Figure 3 shows similar results for various combinations of within- frames analyzed. The recordings included periods that were not
and across-frame differences. Simple combination of difference part of the study; typically recordings began with an empty booth,
images (as illustrated in figure 1) produces a significant improve- depicted a brief flurry of activity as the subject and experimenter
ment over single difference images, but still fails frequently. The entered, and then a relatively long period with the subject seated at
within- and across-frame combination can be applied to curvature the console. So, some fraction of the frames in which no faces were
images as well, and some improvement is seen. But as was the case detected are indeed correct rejections. However, the face detector
for the simple differences, restriction of the search area produces as implemented is trained to detect frontal faces, and fails when the
the largest single improvement; for our test sequence, we obtain head is turned, presenting an oblique or profile view. Similarly, it
near-perfect performance for combination of curvatures of within- fails when the subject tilts his or her head down (for example, to
and across-frame difference images, with a search area restricted to read a clipboard held in the lap). We feel that these misses do not
a neighborhood about the previous location. greatly diminish the value of the dataset, because in these cases a
gaze tracking system would similarly be hard-put to make sense of
the image.
3 Modeling natural head movements
The outputs of the face detector are the location and size of a square
In the previous section, we saw that the eye regions can be tracked window containing the face. From the raw data, we computed the
much more robustly when they are assumed to be relatively near inter-frame displacement of the center of the face-box, for all pairs
to their previous locations. In order to intelligently determine the of temporally contiguous frames in which exactly one face was de-
131
4. References
BARTH , E., Z ETZSCHE , C., AND K RIEGER , G. 1998. Curvature
measures in visual information processing. Open Systems and
Information Dynamics 5, 25–39.
B ROLLY, X. L. C., AND M ULLIGAN , J. B. 2004. Implicit calibra-
tion of a remote gaze tracker. In IEEE Conference on Computer
Vision and Pattern Recognition Workshop (CVPRW), vol. 8, 134.
E BISAWA , Y., AND S ATOH , S. 1993. Effectiveness of pupil area
detection technique using two light sources and image difference
method. In Proceedings of the 15th Annual International Con-
ference of the IEEE Engineering in Medicine and Biology Soci-
ety, San Diego, CA, A. Y. J. Szeto and R. M. Rangayan, Eds.,
1268–1269.
E BISAWA , Y. 2009. Robust pupil detection by image dif-
ference with positional compensation. In Proceedings of the
2009 IEEE International Conference on Virtual Environments,
Human-Computer Interfaces and Measurement Systems, 143–
148.
Figure 4: Plot showing histogram for inter-frame face displace-
ment, computed from a large corpus of video data using the Viola- F ISCHER , U., M C D ONNEL , L., AND O RASANU , J. 2007. Lin-
Jones face detector, as implemented in OpenCV. guistic correlates of team performance: toward a tool for moni-
toring team functioning during space missions. Aviation, Space,
and Environmental Medicine 78, 5 Suppl., B86–B95.
tected in each frame of the pair. Because the detector missed the F OUQUET, J. E., H AVEN , R. E., V ENKATESH , S., AND W EN -
face on some frames when the subject was present, there are gaps STRAND , J. S. 2004. A new optical imaging approach for detect-
in the records, and the number of transitions available for analysis ing drowsy drivers. In IEEE Intelligent Transportation Systems
is somewhat less than the number of frames. The number of tran- Conference, IEEE, WeA6.
sitions analyzed was approximately 47.1% of all transitions, corre-
sponding to 92.4% of the frames in which one face was detected. G RACE , R., B YRNE , V. E., B IERMAN , D. M., L EGRAND , J.,
The results are shown as a histogram of displacements in figure 4. G RICOURT, D., DAVIS , B. K., S TASZEWSKI , J. J., AND C AR -
Face displacements in pixels have been converted to face-widths by NAHAN , B. 2001. A drowsy driver detection system for heavy
dividing each displacement by the average of the box sizes in the vehicles. In Proceedings of the 17th Digital Avionics Systems
two frames making up the pair. Note that the vertical axis is loga- Conference, vol. 2, I36/1–I36/8.
rithmic; the bulk of the displacements are at or near zero, reflecting
the fact that the subjects sat still more than they moved. L IENHART, R., AND M AYDT, J. 2002. An extended set of haar-like
features for rapid object detection. In IEEE ICIP 2002, 900–903.
It can be seen that the central portion of the distribution in figure 4
is within ±1 head-width per second. (Because the voluntary head M ORIMOTO , C. H., KOONS , D., A MIR , A., AND F LICKNER , M.
2000. Pupil detection and tracking using multiple light sources.
movements shown in figures 2 and 3 do not exceed 2 head-widths
Image and Vision Computing 18, 4, 331–335.
per second, we assume that the tails of the distribution shown in
figure 4, which extend beyond the range plotted, represent large T ONG , Y., AND J I , Q. 2008. Automatic eye position detection and
jumps of the face detector to false targets.) Assuming that an eye- tracking under natural facial movement. In Passive Eye Moni-
width is about one fifth of a head-width, this corresponds to 5 eye- toring, Springer, R. I. Hammoud, Ed., 83–107.
widths per second, or 0.083 eye-widths per field of 60 Hz video. For
a high-magnification image in which the eye fills a frame having an V IOLA , P., AND J ONES , M. 2001. Rapid object detection using a
angular extent of 4 x 3 degrees (corresponding to a camera located boosted cascade of simple features. In Proc. IEEE Conference on
at arm’s length from the subject), we see that this corresponds to an Computer Vision and Pattern Recognition (CVPR), vol. 1, 511–
angular rate at the camera of 0.33 degrees per field, or 20 degrees 518.
per second. This provides a performance goal for robotic cameras Z ETZSCHE , C., AND BARTH , E. 1990. Fundamental limits of
used for remote gaze tracking. linear filters in the visual processing of two-dimensional signals.
Vision Res. 30, 1111–1117.
4 Conclusion
In this paper we have demonstrated an novel approach to active il-
lumination eye detection, that combines pairs of difference images
to reduce the effects of head movements. We have also examined
the movement statistics of users working in a naturalistic setting at
a computer console, and found that the majority of movements do
not exceed a speed of 1 head-width per second. This is therefore a
sensible target goal for a head-tracking system designed to keep the
eye region within the area of analysis in a remote gaze-tracking sys-
tem. By considering the camera field-of-view, and average distance
from the subject, we can appropriately size the search region when
repositioning the region-of-interest in a high-resolution image.
132