Selecting a graphical item by pointing with a computer mouse is a ubiquitous task in many graphical user interfaces. Several techniques have been suggested to facilitate this task, for instance, by reducing the required movement distance. Here we
measure the natural coordination of eye and mouse pointer
control across several search and selection tasks. We find that users automatically minimize the distance to likely targets in an intelligent, task dependent way. When target location is highly predictable, top-down knowledge can enable users to initiate pointer movements prior to target fixation. These findings question the utility of existing assistive pointing techniques and suggest that alternative approaches might be more effective.
Digital images can be represented as a two-dimensional matrix of pixels, where each pixel has a discrete location and value. There are three fundamental steps in digital image processing: 1) image acquisition, 2) image enhancement to improve image quality, and 3) image restoration to repair degraded images based on models of image degradation. Additional steps include color image processing, compression, morphological processing, segmentation to separate images into objects, and representation/description to convert raw pixel data into a computer-processable form.
Question & Answer 2.0: A System to facilitate interaction during and after a ...Kai Michael Höver
§ Q&A 2.0 is a system that facilitates interaction during and after lectures.
§ It allows students to ask questions during lectures and continue the discussion afterwards.
§ The document describes existing systems that enable some interaction, but have limitations like not allowing comments to be made in the context of the learning material or only focusing on the time during or after lectures, not both. Q&A 2.0 aims to address these limitations.
This document describes a face recognition technique that uses a hybrid of principal component analysis (PCA) and an artificial neural network. PCA is used to extract global features of the entire face and local features of the eyes, nose, and mouth regions. These features are used as inputs to an artificial neural network for training and testing. The technique aims to leverage both global and local features for face recognition while reducing computation time compared to local-feature-only approaches.
This document discusses optimal image compression techniques based on wavelet transforms. It proposes classifying images as artificial or natural and then applying different wavelet-based compression methods. Haar, Daubechies, Coiflet, and Discrete Meyer wavelet transforms are compared for compressing artificial and natural images. A novel classifier is developed that extracts features from images like entropy, color, and saturation to categorize them. Compression performance and ratio are found to be significantly influenced by first classifying images and then applying the appropriate wavelet technique.
Rotation Invariant Face Recognition using RLBP, LPQ and CONTOURLET TransformIRJET Journal
This document discusses rotation invariant face recognition using three feature extraction techniques: Rotated Local Binary Pattern (RLBP), Local Phase Quantization (LPQ), and Contourlet transform. It first extracts features from input face images using these three techniques. It then applies Linear Discriminant Analysis to reduce the feature dimensions. Finally, it uses k-Nearest Neighbors classification to perform face recognition on the Jaffe dataset. Experimental results show that the face recognition accuracy without LDA is 99.06% and increases to 100% when LDA is used for feature dimension reduction.
A Survey of Image Segmentation based on Artificial Intelligence and Evolution...IOSR Journals
Abstract : In image analysis, segmentation is the partitioning of a digital image into multiple regions (sets of
pixels), according to some homogeneity criterion. The problem of segmentation is a well-studied one in
literature and there are a wide variety of approaches that are used. Different approaches are suited to different
types of images and the quality of output of a particular algorithm is difficult to measure quantitatively due to
the fact that there may be much correct segmentation for a single image. Image segmentation denotes a process
by which a raw input image is partitioned into nonoverlapping regions such that each region is homogeneous
and the union of any two adjacent regions is heterogeneous. A segmented image is considered to be the highest
domain-independent abstraction of an input image. Image segmentation is an important processing step in many
image, video and computer vision applications. Extensive research has been done in creating many different
approaches and algorithms for image segmentation, but it is still difficult to assess whether one algorithm
produces more accurate segmentations than another, whether it be for a particular image or set of images, or
more generally, for a whole class of images.
In this paper, The Survey of Image Segmentation using Artificial Intelligence and Evolutionary Approach
methods that have been proposed in the literature. The rest of the paper is organized as follows. 1.
Introduction, 2.Literature review, 3.Noteworthy contributions in the field of proposed work, 4.Proposed
Methodology, 5.Expected outcome of the proposed research work, 6.Conclusion.
Keywords: Image Segmentation, Segmentation Algorithm, Artificial Intelligence, Evolutionary Algorithm,
Neural Network, Fuzzy Set, Clustering.
Emotion-oriented computing: Possible uses and applicationsAndré Valdestilhas
This article discusses the concepts of using digital television affective computing and computer vision.
The proposal involves the union of some techniques such as capturing facial expressions through a video
camera, use of accelerometers in ball and touch holograms to work a certain level of interactivity with the
viewer. Some uses of the proposal in question are described, such as control of the hearing, background
content, among others. This article reveals numerous benefits that can be addressed with the use of
matters presented which can be applied in a broad context, such as for the blind in video games, among
others
Feature Extraction of Gesture Recognition Based on Image Analysis for Differe...IJERA Editor
Gesture recognition system received great attention in the recent few years because of its manifoldness applications and the ability to interact with machine efficiently through human computer interaction. Gesture is one of human body languages which are popularly used in our daily life. It is a communication system that consists of hand movements and facial expressions via communication by actions and sights. This research mainly focuses on the research of gesture extraction and finger segmentation in the gesture recognition. In this paper, we have used image analysis technologies to create an application by encoding in MATLAB program. We will use this application to segment and extract the finger from one specific gesture. This paper is aimed to give gesture recognition in different natural conditions like dark and glare condition, different distances condition and similar object condition then collect the results to calculate the successful extraction rate.
Digital images can be represented as a two-dimensional matrix of pixels, where each pixel has a discrete location and value. There are three fundamental steps in digital image processing: 1) image acquisition, 2) image enhancement to improve image quality, and 3) image restoration to repair degraded images based on models of image degradation. Additional steps include color image processing, compression, morphological processing, segmentation to separate images into objects, and representation/description to convert raw pixel data into a computer-processable form.
Question & Answer 2.0: A System to facilitate interaction during and after a ...Kai Michael Höver
§ Q&A 2.0 is a system that facilitates interaction during and after lectures.
§ It allows students to ask questions during lectures and continue the discussion afterwards.
§ The document describes existing systems that enable some interaction, but have limitations like not allowing comments to be made in the context of the learning material or only focusing on the time during or after lectures, not both. Q&A 2.0 aims to address these limitations.
This document describes a face recognition technique that uses a hybrid of principal component analysis (PCA) and an artificial neural network. PCA is used to extract global features of the entire face and local features of the eyes, nose, and mouth regions. These features are used as inputs to an artificial neural network for training and testing. The technique aims to leverage both global and local features for face recognition while reducing computation time compared to local-feature-only approaches.
This document discusses optimal image compression techniques based on wavelet transforms. It proposes classifying images as artificial or natural and then applying different wavelet-based compression methods. Haar, Daubechies, Coiflet, and Discrete Meyer wavelet transforms are compared for compressing artificial and natural images. A novel classifier is developed that extracts features from images like entropy, color, and saturation to categorize them. Compression performance and ratio are found to be significantly influenced by first classifying images and then applying the appropriate wavelet technique.
Rotation Invariant Face Recognition using RLBP, LPQ and CONTOURLET TransformIRJET Journal
This document discusses rotation invariant face recognition using three feature extraction techniques: Rotated Local Binary Pattern (RLBP), Local Phase Quantization (LPQ), and Contourlet transform. It first extracts features from input face images using these three techniques. It then applies Linear Discriminant Analysis to reduce the feature dimensions. Finally, it uses k-Nearest Neighbors classification to perform face recognition on the Jaffe dataset. Experimental results show that the face recognition accuracy without LDA is 99.06% and increases to 100% when LDA is used for feature dimension reduction.
A Survey of Image Segmentation based on Artificial Intelligence and Evolution...IOSR Journals
Abstract : In image analysis, segmentation is the partitioning of a digital image into multiple regions (sets of
pixels), according to some homogeneity criterion. The problem of segmentation is a well-studied one in
literature and there are a wide variety of approaches that are used. Different approaches are suited to different
types of images and the quality of output of a particular algorithm is difficult to measure quantitatively due to
the fact that there may be much correct segmentation for a single image. Image segmentation denotes a process
by which a raw input image is partitioned into nonoverlapping regions such that each region is homogeneous
and the union of any two adjacent regions is heterogeneous. A segmented image is considered to be the highest
domain-independent abstraction of an input image. Image segmentation is an important processing step in many
image, video and computer vision applications. Extensive research has been done in creating many different
approaches and algorithms for image segmentation, but it is still difficult to assess whether one algorithm
produces more accurate segmentations than another, whether it be for a particular image or set of images, or
more generally, for a whole class of images.
In this paper, The Survey of Image Segmentation using Artificial Intelligence and Evolutionary Approach
methods that have been proposed in the literature. The rest of the paper is organized as follows. 1.
Introduction, 2.Literature review, 3.Noteworthy contributions in the field of proposed work, 4.Proposed
Methodology, 5.Expected outcome of the proposed research work, 6.Conclusion.
Keywords: Image Segmentation, Segmentation Algorithm, Artificial Intelligence, Evolutionary Algorithm,
Neural Network, Fuzzy Set, Clustering.
Emotion-oriented computing: Possible uses and applicationsAndré Valdestilhas
This article discusses the concepts of using digital television affective computing and computer vision.
The proposal involves the union of some techniques such as capturing facial expressions through a video
camera, use of accelerometers in ball and touch holograms to work a certain level of interactivity with the
viewer. Some uses of the proposal in question are described, such as control of the hearing, background
content, among others. This article reveals numerous benefits that can be addressed with the use of
matters presented which can be applied in a broad context, such as for the blind in video games, among
others
Feature Extraction of Gesture Recognition Based on Image Analysis for Differe...IJERA Editor
Gesture recognition system received great attention in the recent few years because of its manifoldness applications and the ability to interact with machine efficiently through human computer interaction. Gesture is one of human body languages which are popularly used in our daily life. It is a communication system that consists of hand movements and facial expressions via communication by actions and sights. This research mainly focuses on the research of gesture extraction and finger segmentation in the gesture recognition. In this paper, we have used image analysis technologies to create an application by encoding in MATLAB program. We will use this application to segment and extract the finger from one specific gesture. This paper is aimed to give gesture recognition in different natural conditions like dark and glare condition, different distances condition and similar object condition then collect the results to calculate the successful extraction rate.
This document provides an exploratory review of soft computing techniques for image segmentation. It discusses various segmentation techniques including discontinuity-based techniques like point, line and edge detection using spatial filtering. Thresholding techniques like global, adaptive and multi-level thresholding are also covered. Region-based techniques such as region growing, region splitting/merging and morphological watersheds are summarized. The document concludes that future work can focus on developing genetic segmentation filters using a genetic algorithm approach for medical image segmentation.
This paper examines single gaze gestures (SGGs) as a selection method for gaze-controlled interfaces. SGGs involve making a single point-to-point eye movement between two on-screen locations. The study evaluated horizontal and vertical, long and short SGGs on two eye tracking devices. It found that long SGGs took significantly longer to complete than short SGGs. Horizontal SGGs were also completed significantly faster than vertical SGGs. However, there was no significant difference in error rates between horizontal and vertical SGGs. The study provides evidence that SGGs can be an effective selection technique, with properties like selection time varying based on gesture length and direction.
Pattern Approximation Based Generalized Image Noise Reduction Using Adaptive ...IJECEIAES
The problem of noise interference with the image always occurs irrespective of whatever precaution is taken. Challenging issues with noise reduction are diversity of characteristics involved with source of noise and in result; it is difficult to develop a universal solution. This paper has proposed neural network based generalize solution of noise reduction by mapping the problem as pattern approximation. Considering the statistical relationship among local region pixels in the noise free image as normal patterns, feedforward neural network is applied to acquire the knowledge available within such patterns. Adaptiveness is applied in the slope of transfer function to improve the learning process. Acquired normal patterns knowledge is utilized to reduce the level of different type of noise available within an image by recorrection of noisy patterns through pattern approximation. The proposed restoration method does not need any estimation of noise model characteristics available in the image not only that it can reduce the mixer of different types of noise efficiently. The proposed method has high processing speed along with simplicity in design. Restoration of gray scale image as well as color image has done, which has suffered from different types of noise like, Gaussian noise, salt &peper, speckle noise and mixer of it.
IRJET- Kidney Stone Classification using Deep Neural Networks and Facilitatin...IRJET Journal
The document discusses classifying kidney stone images using deep neural networks and facilitating diagnosis using IoT. Kidney stone images are acquired and preprocessed by converting to grayscale, enhancing, and segmenting the area of interest. Texture features are extracted using active contour segmentation and classified using a deep neural network model. The results, including stone type and treatment recommendations, are sent to the cloud where doctors and patients can access them, allowing automated diagnosis without human intervention.
This document summarizes a research paper on techniques for binarizing degraded document images. It discusses how degraded documents often have mixed foreground and background pixels that need to be separated. The proposed method uses contrast adjustment, grey scale edge detection, thresholding and post-processing to binarize degraded images. It first inverts the image contrast, then uses grey scale detection to find text stroke edges. Pixels are classified and thresholding is used to create a binary image. Post-processing removes background pixels to output a clean image with only text strokes. The method is tested on degraded novel and book images and produces separated, readable text from the backgrounds.
Improving image resolution through the cra algorithm involved recycling proce...csandit
Image processing concepts are widely used in medical fields. Digital images are prone to a
variety of types of noise. Noise is the result of errors in the image acquisition process for
reconstruction that result in pixel values that reflect the true intensities of the real scenes. A lot
of researchers are working on the field analysis and processing of multi-dimensional images.
Work previously hasn’t sufficient to stop them, so they continue performance work is due by the
researcher. In this paper we contribute a novel research work for analysis and performance
improvement about to image resolution. We proposed Concede Reconstruction Algorithm (CRA)
Involved Recycling Process to reduce the remained problem in improvement part of an image
processing. The CRA algorithms have better response from researcher to use them
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...Kalle
To estimate viewer’s contextual understanding, features of their
eye-movements while viewing question statements in response to definition statements, and features of correct and incorrect responses were extracted and compared. Twelve directional features
of eye-movements across a two-dimensional space were created, and these features were compared between correct and incorrect responses. The procedure of estimating the response was developed with Support Vector Machines, using these features. The estimation performance and accuracy were assessed across combinations of features. The number of definition statements, which needed to be memorized to answer the question statements during the experiment, affected the estimation accuracy. These results provide evidence that features of eye-movements during reading statements
can be used as an index of contextual understanding.
IRJET- Comparative Study of Artificial Neural Networks and Convolutional N...IRJET Journal
This document discusses and compares artificial neural networks and convolutional neural networks for crop disease detection using images. It first provides background on the importance of early crop disease detection in India. It then describes the image preprocessing, segmentation, and feature extraction steps involved, including converting to HSV color space and extracting texture features using GLCM. Artificial neural networks and convolutional neural networks are introduced for classification. The document conducts a literature review on previous work related to image preprocessing techniques, segmentation algorithms like K-means clustering, and feature extraction methods. In summary, it analyzes the process of detecting crop diseases from images using machine learning techniques.
Exploring visual and motion saliency for automatic video object extractionMuthu Samy
Sybian Technologies Pvt Ltd
Final Year Projects & Real Time live Projects
JAVA(All Domains)
DOTNET(All Domains)
ANDROID
EMBEDDED
VLSI
MATLAB
Project Support
Abstract, Diagrams, Review Details, Relevant Materials, Presentation,
Supporting Documents, Software E-Books,
Software Development Standards & Procedure
E-Book, Theory Classes, Lab Working Programs, Project Design & Implementation
24/7 lab session
Final Year Projects For BE,ME,B.Sc,M.Sc,B.Tech,BCA,MCA
PROJECT DOMAIN:
Cloud Computing
Networking
Network Security
PARALLEL AND DISTRIBUTED SYSTEM
Data Mining
Mobile Computing
Service Computing
Software Engineering
Image Processing
Bio Medical / Medical Imaging
Contact Details:
Sybian Technologies Pvt Ltd,
No,33/10 Meenakshi Sundaram Building,
Sivaji Street,
(Near T.nagar Bus Terminus)
T.Nagar,
Chennai-600 017
Ph:044 42070551
Mobile No:9790877889,9003254624,7708845605
Mail Id:sybianprojects@gmail.com,sunbeamvijay@yahoo.com
Maximizing Strength of Digital Watermarks Using Fuzzy Logicsipij
In this paper, we propose a novel digital watermarking scheme in DCT domain based fuzzy inference system and the human visual system to adapt the embedding strength of different blocks. Firstly, the original image is divided into some 8×8 blocks, and then fuzzy inference system according to different textural features and luminance of each block decide adaptively different embedding strengths. The watermark detection adopts correlation technology. Experimental results show that the proposed scheme has good imperceptibility and high robustness to common image processing operators.
IRJET - Deep Learning Approach to Inpainting and Outpainting SystemIRJET Journal
This document discusses a deep learning approach for image inpainting and outpainting. It proposes a new generative model-based approach using a fully convolutional neural network that can process images with multiple holes at variable locations and sizes. The model aims to not only synthesize novel image structures, but also explicitly utilize surrounding image features as references during training to generate better predictions. Experiments on faces, textures and natural images demonstrate the proposed approach generates higher quality inpainting results than existing methods. It aims to address limitations of CNNs in borrowing information from distant areas by leveraging texture and patch synthesis approaches.
This document discusses a hand gesture recognition system for underprivileged individuals. It begins by outlining the key steps in hand gesture recognition systems: image capture, pre-processing, segmentation, feature extraction and gesture recognition. It then goes into more detail on specific techniques for each step, such as thresholding and edge detection for segmentation. The document also covers applications like access control, sign language translation and future areas like biometric authentication. In conclusion, it proposes that hand gesture recognition can help disabled individuals communicate through accessible human-computer interaction.
Image Enhancement by Image Fusion for Crime InvestigationCSCJournals
This document proposes a method for image enhancement through image fusion for crime investigation applications. It summarizes existing image enhancement techniques like histogram equalization and presents their limitations. It then describes the proposed method which involves constructing an image pyramid and performing a wavelet transformation on input images. The pyramid and wavelet transformed images are then fused to generate an enhanced output image with improved contrast and information content. Experimental results on a surveillance camera image show that the proposed fusion scheme provides better perception for human visual analysis compared to traditional enhancement techniques.
This document provides an overview and analysis of various image smoothing techniques. It begins with an introduction to image smoothing and its importance in digital image processing. Then, it analyzes several common image smoothing algorithms in detail, including Gaussian smoothing, edge-preserved filtering, bilateral filtering, optimization-based image filtering, non-linear diffusion filtering, robust smoothing filtering, gradient weighting filtering, and guided image filtering. For each algorithm, it explains the underlying concepts and mathematical principles. It finds that appropriate choice of smoothing techniques depends on factors like the imaging modality, noise characteristics, and task requirements. The document aims to provide insights into widely used image smoothing approaches.
IRJET - Effective Workflow for High-Performance Recognition of Fruits using M...IRJET Journal
This document presents research on fruit recognition using machine learning approaches. The researchers used the fruit-360 dataset containing 74,572 images of 109 fruit classes. They applied feature extraction techniques including HU moments, Haralick texture, and color histogram. Several machine learning classifiers were then trained on the extracted features, including decision tree, K-nearest neighbors, linear discriminant analysis, logistic regression, naive Bayes, random forest, and support vector machine. The models were evaluated using metrics like sensitivity, specificity, precision, F1-score, and accuracy. The results found that K-nearest neighbors and random forest classifiers achieved the best performance with a false positive rate of 0% and high accuracy, outperforming previous fruit recognition studies.
A Framework for Human Action Detection via Extraction of Multimodal FeaturesCSCJournals
This work discusses the application of an Artificial Intelligence technique called data extraction and a process-based ontology in constructing experimental qualitative models for video retrieval and detection. We present a framework architecture that uses multimodality features as the knowledge representation scheme to model the behaviors of a number of human actions in the video scenes. The main focus of this paper placed on the design of two main components (model classifier and inference engine) for a tool abbreviated as VASD (Video Action Scene Detector) for retrieving and detecting human actions from video scenes. The discussion starts by presenting the workflow of the retrieving and detection process and the automated model classifier construction logic. We then move on to demonstrate how the constructed classifiers can be used with multimodality features for detecting human actions. Finally, behavioral explanation manifestation is discussed. The simulator is implemented in bilingual; Math Lab and C++ are at the backend supplying data and theories while Java handles all front-end GUI and action pattern updating. To compare the usefulness of the proposed framework, several experiments were conducted and the results were obtained by using visual features only (77.89% for precision; 72.10% for recall), audio features only (62.52% for precision; 48.93% for recall) and combined audiovisual (90.35% for precision; 90.65% for recall).
THE EFFECT OF IMPLEMENTING OF NONLINEAR FILTERS FOR ENHANCING MEDICAL IMAGES ...ijcsit
Although this huge development in medical imaging tools, we find that there are some human mistakes in the process of filming medical images, where some errors result in distortions in the image and change some medical image properties which affect the disease diagnosis correctly.Medical images are one of the fundamental images, because they are used in the most sensitive field which is a medical field. The
objective of the study is to identify the effect of implement non-linear filters in enhancing medical images,using the strongest and most popular program MATLAB, and because of its advantages in image processing. After implementation the researcher concluded that we will get the best result for medical image enhancement by using median filter which is one of the non-linear filters,non-linear filters implemented using Matlab functions
This document presents a method for image upscaling using a fuzzy ARTMAP neural network. It begins with an introduction to image upscaling and interpolation techniques. It then provides background on ARTMAP neural networks and fuzzy logic. The proposed method uses a linear interpolation algorithm trained with an ARTMAP network. Results show the method performs better than nearest neighbor interpolation in terms of peak signal-to-noise ratio, mean squared error, and structural similarity, though not as high as bicubic interpolation. Overall, the fuzzy ARTMAP network provides an effective way to perform image upscaling with fewer artifacts than traditional methods.
What can Leeds-based creative agency, talktojason do for youJason Kelly
This document outlines various public relations and communications services including media training, conferences, interviews, coverage, opinion pieces, releases, consultancy, brand architecture, buzz generation, consumer engagement, communications channels, marketing literature, and digital creative. It repeatedly directs the reader to talk to Jason for further details on these services.
The document summarizes various remote and local attacks that can be performed against Microsoft Windows operating systems. It discusses vulnerabilities in Windows services like RPC, SMB, and LSASS that can allow remote code execution. It also covers techniques like token kidnapping that can be used to elevate privileges on Windows systems. The document demonstrates exploits for the MS09-050 SMB vulnerability and KiTrap0d local kernel exploit. It warns that applications like Internet Explorer and Office macros can also be weaponized to attack Windows machines. Throughout it emphasizes that Windows OSes are prone to security issues.
This document provides an exploratory review of soft computing techniques for image segmentation. It discusses various segmentation techniques including discontinuity-based techniques like point, line and edge detection using spatial filtering. Thresholding techniques like global, adaptive and multi-level thresholding are also covered. Region-based techniques such as region growing, region splitting/merging and morphological watersheds are summarized. The document concludes that future work can focus on developing genetic segmentation filters using a genetic algorithm approach for medical image segmentation.
This paper examines single gaze gestures (SGGs) as a selection method for gaze-controlled interfaces. SGGs involve making a single point-to-point eye movement between two on-screen locations. The study evaluated horizontal and vertical, long and short SGGs on two eye tracking devices. It found that long SGGs took significantly longer to complete than short SGGs. Horizontal SGGs were also completed significantly faster than vertical SGGs. However, there was no significant difference in error rates between horizontal and vertical SGGs. The study provides evidence that SGGs can be an effective selection technique, with properties like selection time varying based on gesture length and direction.
Pattern Approximation Based Generalized Image Noise Reduction Using Adaptive ...IJECEIAES
The problem of noise interference with the image always occurs irrespective of whatever precaution is taken. Challenging issues with noise reduction are diversity of characteristics involved with source of noise and in result; it is difficult to develop a universal solution. This paper has proposed neural network based generalize solution of noise reduction by mapping the problem as pattern approximation. Considering the statistical relationship among local region pixels in the noise free image as normal patterns, feedforward neural network is applied to acquire the knowledge available within such patterns. Adaptiveness is applied in the slope of transfer function to improve the learning process. Acquired normal patterns knowledge is utilized to reduce the level of different type of noise available within an image by recorrection of noisy patterns through pattern approximation. The proposed restoration method does not need any estimation of noise model characteristics available in the image not only that it can reduce the mixer of different types of noise efficiently. The proposed method has high processing speed along with simplicity in design. Restoration of gray scale image as well as color image has done, which has suffered from different types of noise like, Gaussian noise, salt &peper, speckle noise and mixer of it.
IRJET- Kidney Stone Classification using Deep Neural Networks and Facilitatin...IRJET Journal
The document discusses classifying kidney stone images using deep neural networks and facilitating diagnosis using IoT. Kidney stone images are acquired and preprocessed by converting to grayscale, enhancing, and segmenting the area of interest. Texture features are extracted using active contour segmentation and classified using a deep neural network model. The results, including stone type and treatment recommendations, are sent to the cloud where doctors and patients can access them, allowing automated diagnosis without human intervention.
This document summarizes a research paper on techniques for binarizing degraded document images. It discusses how degraded documents often have mixed foreground and background pixels that need to be separated. The proposed method uses contrast adjustment, grey scale edge detection, thresholding and post-processing to binarize degraded images. It first inverts the image contrast, then uses grey scale detection to find text stroke edges. Pixels are classified and thresholding is used to create a binary image. Post-processing removes background pixels to output a clean image with only text strokes. The method is tested on degraded novel and book images and produces separated, readable text from the backgrounds.
Improving image resolution through the cra algorithm involved recycling proce...csandit
Image processing concepts are widely used in medical fields. Digital images are prone to a
variety of types of noise. Noise is the result of errors in the image acquisition process for
reconstruction that result in pixel values that reflect the true intensities of the real scenes. A lot
of researchers are working on the field analysis and processing of multi-dimensional images.
Work previously hasn’t sufficient to stop them, so they continue performance work is due by the
researcher. In this paper we contribute a novel research work for analysis and performance
improvement about to image resolution. We proposed Concede Reconstruction Algorithm (CRA)
Involved Recycling Process to reduce the remained problem in improvement part of an image
processing. The CRA algorithms have better response from researcher to use them
Nakayama Estimation Of Viewers Response For Contextual Understanding Of Tasks...Kalle
To estimate viewer’s contextual understanding, features of their
eye-movements while viewing question statements in response to definition statements, and features of correct and incorrect responses were extracted and compared. Twelve directional features
of eye-movements across a two-dimensional space were created, and these features were compared between correct and incorrect responses. The procedure of estimating the response was developed with Support Vector Machines, using these features. The estimation performance and accuracy were assessed across combinations of features. The number of definition statements, which needed to be memorized to answer the question statements during the experiment, affected the estimation accuracy. These results provide evidence that features of eye-movements during reading statements
can be used as an index of contextual understanding.
IRJET- Comparative Study of Artificial Neural Networks and Convolutional N...IRJET Journal
This document discusses and compares artificial neural networks and convolutional neural networks for crop disease detection using images. It first provides background on the importance of early crop disease detection in India. It then describes the image preprocessing, segmentation, and feature extraction steps involved, including converting to HSV color space and extracting texture features using GLCM. Artificial neural networks and convolutional neural networks are introduced for classification. The document conducts a literature review on previous work related to image preprocessing techniques, segmentation algorithms like K-means clustering, and feature extraction methods. In summary, it analyzes the process of detecting crop diseases from images using machine learning techniques.
Exploring visual and motion saliency for automatic video object extractionMuthu Samy
Sybian Technologies Pvt Ltd
Final Year Projects & Real Time live Projects
JAVA(All Domains)
DOTNET(All Domains)
ANDROID
EMBEDDED
VLSI
MATLAB
Project Support
Abstract, Diagrams, Review Details, Relevant Materials, Presentation,
Supporting Documents, Software E-Books,
Software Development Standards & Procedure
E-Book, Theory Classes, Lab Working Programs, Project Design & Implementation
24/7 lab session
Final Year Projects For BE,ME,B.Sc,M.Sc,B.Tech,BCA,MCA
PROJECT DOMAIN:
Cloud Computing
Networking
Network Security
PARALLEL AND DISTRIBUTED SYSTEM
Data Mining
Mobile Computing
Service Computing
Software Engineering
Image Processing
Bio Medical / Medical Imaging
Contact Details:
Sybian Technologies Pvt Ltd,
No,33/10 Meenakshi Sundaram Building,
Sivaji Street,
(Near T.nagar Bus Terminus)
T.Nagar,
Chennai-600 017
Ph:044 42070551
Mobile No:9790877889,9003254624,7708845605
Mail Id:sybianprojects@gmail.com,sunbeamvijay@yahoo.com
Maximizing Strength of Digital Watermarks Using Fuzzy Logicsipij
In this paper, we propose a novel digital watermarking scheme in DCT domain based fuzzy inference system and the human visual system to adapt the embedding strength of different blocks. Firstly, the original image is divided into some 8×8 blocks, and then fuzzy inference system according to different textural features and luminance of each block decide adaptively different embedding strengths. The watermark detection adopts correlation technology. Experimental results show that the proposed scheme has good imperceptibility and high robustness to common image processing operators.
IRJET - Deep Learning Approach to Inpainting and Outpainting SystemIRJET Journal
This document discusses a deep learning approach for image inpainting and outpainting. It proposes a new generative model-based approach using a fully convolutional neural network that can process images with multiple holes at variable locations and sizes. The model aims to not only synthesize novel image structures, but also explicitly utilize surrounding image features as references during training to generate better predictions. Experiments on faces, textures and natural images demonstrate the proposed approach generates higher quality inpainting results than existing methods. It aims to address limitations of CNNs in borrowing information from distant areas by leveraging texture and patch synthesis approaches.
This document discusses a hand gesture recognition system for underprivileged individuals. It begins by outlining the key steps in hand gesture recognition systems: image capture, pre-processing, segmentation, feature extraction and gesture recognition. It then goes into more detail on specific techniques for each step, such as thresholding and edge detection for segmentation. The document also covers applications like access control, sign language translation and future areas like biometric authentication. In conclusion, it proposes that hand gesture recognition can help disabled individuals communicate through accessible human-computer interaction.
Image Enhancement by Image Fusion for Crime InvestigationCSCJournals
This document proposes a method for image enhancement through image fusion for crime investigation applications. It summarizes existing image enhancement techniques like histogram equalization and presents their limitations. It then describes the proposed method which involves constructing an image pyramid and performing a wavelet transformation on input images. The pyramid and wavelet transformed images are then fused to generate an enhanced output image with improved contrast and information content. Experimental results on a surveillance camera image show that the proposed fusion scheme provides better perception for human visual analysis compared to traditional enhancement techniques.
This document provides an overview and analysis of various image smoothing techniques. It begins with an introduction to image smoothing and its importance in digital image processing. Then, it analyzes several common image smoothing algorithms in detail, including Gaussian smoothing, edge-preserved filtering, bilateral filtering, optimization-based image filtering, non-linear diffusion filtering, robust smoothing filtering, gradient weighting filtering, and guided image filtering. For each algorithm, it explains the underlying concepts and mathematical principles. It finds that appropriate choice of smoothing techniques depends on factors like the imaging modality, noise characteristics, and task requirements. The document aims to provide insights into widely used image smoothing approaches.
IRJET - Effective Workflow for High-Performance Recognition of Fruits using M...IRJET Journal
This document presents research on fruit recognition using machine learning approaches. The researchers used the fruit-360 dataset containing 74,572 images of 109 fruit classes. They applied feature extraction techniques including HU moments, Haralick texture, and color histogram. Several machine learning classifiers were then trained on the extracted features, including decision tree, K-nearest neighbors, linear discriminant analysis, logistic regression, naive Bayes, random forest, and support vector machine. The models were evaluated using metrics like sensitivity, specificity, precision, F1-score, and accuracy. The results found that K-nearest neighbors and random forest classifiers achieved the best performance with a false positive rate of 0% and high accuracy, outperforming previous fruit recognition studies.
A Framework for Human Action Detection via Extraction of Multimodal FeaturesCSCJournals
This work discusses the application of an Artificial Intelligence technique called data extraction and a process-based ontology in constructing experimental qualitative models for video retrieval and detection. We present a framework architecture that uses multimodality features as the knowledge representation scheme to model the behaviors of a number of human actions in the video scenes. The main focus of this paper placed on the design of two main components (model classifier and inference engine) for a tool abbreviated as VASD (Video Action Scene Detector) for retrieving and detecting human actions from video scenes. The discussion starts by presenting the workflow of the retrieving and detection process and the automated model classifier construction logic. We then move on to demonstrate how the constructed classifiers can be used with multimodality features for detecting human actions. Finally, behavioral explanation manifestation is discussed. The simulator is implemented in bilingual; Math Lab and C++ are at the backend supplying data and theories while Java handles all front-end GUI and action pattern updating. To compare the usefulness of the proposed framework, several experiments were conducted and the results were obtained by using visual features only (77.89% for precision; 72.10% for recall), audio features only (62.52% for precision; 48.93% for recall) and combined audiovisual (90.35% for precision; 90.65% for recall).
THE EFFECT OF IMPLEMENTING OF NONLINEAR FILTERS FOR ENHANCING MEDICAL IMAGES ...ijcsit
Although this huge development in medical imaging tools, we find that there are some human mistakes in the process of filming medical images, where some errors result in distortions in the image and change some medical image properties which affect the disease diagnosis correctly.Medical images are one of the fundamental images, because they are used in the most sensitive field which is a medical field. The
objective of the study is to identify the effect of implement non-linear filters in enhancing medical images,using the strongest and most popular program MATLAB, and because of its advantages in image processing. After implementation the researcher concluded that we will get the best result for medical image enhancement by using median filter which is one of the non-linear filters,non-linear filters implemented using Matlab functions
This document presents a method for image upscaling using a fuzzy ARTMAP neural network. It begins with an introduction to image upscaling and interpolation techniques. It then provides background on ARTMAP neural networks and fuzzy logic. The proposed method uses a linear interpolation algorithm trained with an ARTMAP network. Results show the method performs better than nearest neighbor interpolation in terms of peak signal-to-noise ratio, mean squared error, and structural similarity, though not as high as bicubic interpolation. Overall, the fuzzy ARTMAP network provides an effective way to perform image upscaling with fewer artifacts than traditional methods.
What can Leeds-based creative agency, talktojason do for youJason Kelly
This document outlines various public relations and communications services including media training, conferences, interviews, coverage, opinion pieces, releases, consultancy, brand architecture, buzz generation, consumer engagement, communications channels, marketing literature, and digital creative. It repeatedly directs the reader to talk to Jason for further details on these services.
The document summarizes various remote and local attacks that can be performed against Microsoft Windows operating systems. It discusses vulnerabilities in Windows services like RPC, SMB, and LSASS that can allow remote code execution. It also covers techniques like token kidnapping that can be used to elevate privileges on Windows systems. The document demonstrates exploits for the MS09-050 SMB vulnerability and KiTrap0d local kernel exploit. It warns that applications like Internet Explorer and Office macros can also be weaponized to attack Windows machines. Throughout it emphasizes that Windows OSes are prone to security issues.
Sam Nathews, age 24, is applying for PR News People to Watch – Rising PR Stars 30 & Under. The supporting materials include clips from various media outlets about his work on media relations for the Attorney General and an iDECIDE youth voting campaign featuring Bella Thorne and Shaquille O'Neal that generated coverage in publications like Daily Mail, Us Weekly, and Just Jared Jr. Social media support for the campaign came from Bella Thorne and Shaquille O'Neal.
The document discusses several album cover designs and evaluates their suitability for inspiration for designing a cover for the band "Fire Drive Tiger". Key points:
- The document analyzes 9 different album covers, commenting on colors, images, layouts, and how well they represent the music genre.
- Covers noted as potentially inspiring include ones with fiery colors, cut-out cartoon-like images of band members, and borders.
- The goal is to create a cover that is artistic, cartoon-like, fun yet rock-and-roll style, to represent Fire Drive Tiger's music.
- The covers of The Ting Tings, Rihanna, and Blur are identified as most inspiring
La Unión Europea ha acordado un embargo petrolero contra Rusia en respuesta a su invasión de Ucrania. El embargo forma parte de un sexto paquete de sanciones y prohibirá la mayoría de las importaciones de petróleo ruso en la UE a finales de este año. Alemania y Polonia han acordado dejar de importar petróleo ruso a través de oleoductos para finales de 2022 como parte de este embargo.
Southern Electronics (Bangalore) Private Limited is an avionics company located in Bangalore, India. It has developed several avionics systems including a fire detection and warning system, high energy ignition unit, standby centralised warning panel, intercom systems, digital clocks, illuminated lighting panels, and fire warning control boxes. It also provides test and ground support equipment, designs switches, and has infrastructure like humidity and temperature chambers for testing. Southern Electronics works with major defense organizations in India like Hindustan Aeronautics Limited and has design organization approvals to certify its avionics systems.
Minkälaisia apuja nykyaikaiset työkalut tuovat digitaaliseen markkinointiin? Miten markkinoinnin rutiinitoimenpiteitä voidaan automatisoida? 18.10.2012 DiViA-foorumissa pidetty esitys nykyaikaisesta, personoidusta asiakaskokemuksesta.
Dokumen ini berisi soalan ujian untuk persatuan kemahiran hidup di sekolah menengah. Soalan-soalan meliputi identiti penasihat dan pengerusi persatuan, jumlah ahli, aktiviti yang dilakukan, serta cadangan aktiviti masa depan. Juga terdapat soalan pilihan berganda mengenai makanan, kebersihan, pakaian, dan lainnya.
This document discusses configuring, customizing, and troubleshooting Outlook Express. It provides instructions on setting up email and newsgroup accounts in Outlook Express, including configuring account properties and servers. Potential issues that could arise are also addressed, such as being unable to send/receive email or access newsgroups. The document also covers customizing the Outlook Express interface and resolving related issues. Key topics covered include email/newsgroup accounts, account configuration, troubleshooting connectivity problems, and customizing the Outlook Express view.
This paper introduce practical techniques used by hackers to break the wireless security.
We recommend that the reader should have basic knowledge of wireless operation.
El documento describe las cuatro estaciones del año - invierno, primavera, otoño y verano - y las condiciones meteorológicas típicas de cada una, incluyendo nieve, sol, lluvia, calor, viento y frío.
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Waqas Tariq
This document describes a study that evaluated the Microsoft Kinect device for natural interaction in an information visualization system called MetricSPlat. The researchers hypothesized that Kinect would enable more efficient interaction than a mouse for tasks like identifying clusters and outliers in multidimensional data projections. They used a participatory design process with users to develop an interaction scheme for controlling MetricSPlat with Kinect gestures. Usability tests were conducted during design to evaluate each iteration. After finalizing the Kinect scheme, comparative usability tests were performed between Kinect and mouse. The results found that while users reported high satisfaction with Kinect, it was less efficient than the mouse in terms of task completion times and precision for the specific visualization tasks in the
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Waqas Tariq
We verify the hypothesis that Microsoft’s Kinect device is tailored for defining more efficient interaction compared to the commodity mouse device in the context of information visualization. For this goal, we used Kinect during interaction design and evaluation considering an application on information visualization (over agrometeorological, cars, and flowers datasets). The devices were tested over a visualization technique based on clouds of points (multidimensional projection) that can be manipulated by rotation, scaling, and translation. The design was carried according to technique Participatory Design (ISO 13407) and the evaluation answered to a vast set of Usability Tests. In the tests, the users reported high satisfaction scores (easiness and preference) but, also, they signed out with low efficiency scores (time and precision). In the specific context of a multidimensional-projection visualization, our conclusion is that, in respect to user acceptance, Kinect is a device adequate for natural interaction; but, for desktop-based production, it still cannot compete with the traditional long-term mouse design.
Design and Evaluation Case Study: Evaluating The Kinect Device In The Task of...Waqas Tariq
We verify the hypothesis that Microsoft’s Kinect device is tailored for defining more efficient interaction compared to the commodity mouse device in the context of information visualization. For this goal, we used Kinect during interaction design and evaluation considering an application on information visualization (over agrometeorological, cars, and flowers datasets). The devices were tested over a visualization technique based on clouds of points (multidimensional projection) that can be manipulated by rotation, scaling, and translation. The design was carried according to technique Participatory Design (ISO 13407) and the evaluation answered to a vast set of Usability Tests. In the tests, the users reported high satisfaction scores (easiness and preference) but, also, they signed out with low efficiency scores (time and precision). In the specific context of a multidimensional-projection visualization, our conclusion is that, in respect to user acceptance, Kinect is a device adequate for natural interaction; but, for desktop-based production, it still cannot compete with the traditional long-term mouse design.
Computer vision for interactive computer graphicsShah Alam Sabuj
This document discusses computer vision and its uses for interactive computer graphics. Computer vision involves acquiring, processing, analyzing and understanding images from the real world in order to produce information. It allows computers to interpret user movements and gestures through algorithms like tracking, shape recognition and motion analysis. These visual algorithms enable interactive applications where the computer can track and respond to both large and small objects in real-time, creating new possibilities for human-computer interaction. Examples provided demonstrate how computer vision can be used to track a user's hand gestures to control a television interface.
Lift using projected coded light for finger tracking and device augmentationShang Ma
We present Lift, a visible light-enabled finger
tracking and object localization technique that allows users to
perform freestyle multi-touch gestures on any object’s surface in
an everyday environment. By projecting encoded visible patterns
onto an object’s surface (e.g. paper, display, or table), and
localizing the user’s fingers with light sensors, Lift offers users a
richer interactive space than the device’s existing interfaces.
Additionally, everyday objects can be augmented by attaching
sensor units onto their surface to accept multi-touch gesture
input. We also present two applications as proof of concept.
Finally, results from our experiments indicate that Lift can
localize ten fingers simultaneously with an average accuracy of
1.7 millimeter and an average refresh rate of 84 Hz with 31
milliseconds delay on WiFi and 23 milliseconds delay on serial
communication, making gesture recognition on non-
instrumented objects possible.
DragGAN is a new AI image editing tool that lets you manipulate images with simple drag controls developed by researchers at the University of California, Berkeley.
It uses generative AI to create realistic changes to the structure and appearance of objects in images. You can also rotate images as if they were 3D models.
The user can then use a drag-and-drop interface to edit the image. DragGAN will then generate a new image that reflects the user's edits.
This document provides an introduction and overview of a project on vision-based hand gesture recognition. It discusses the motivation for the project and how hand gestures can provide a more natural human-computer interaction compared to traditional input devices like keyboards and mice. The document outlines the objectives of the project, which are to develop a system that can identify specific hand gestures using a webcam and interpret them to control mouse operations on a computer. It also provides an overview of the organization of the project report and the topics that will be discussed in subsequent chapters, such as the literature review, proposed methodology, results, and conclusions.
1. The document summarizes a student project that aims to create a virtual try-on application using augmented reality. It surveys existing methods for tasks like clothing segmentation, human pose estimation, and virtual try-on that could be used to build the application.
2. It discusses approaches the students investigated like using depth cameras for measurements, non-depth based methods using computer vision, parsing clothes and humans, and existing work on 2D virtual try-on.
3. The students implemented initial modules for their pipeline including a U-Net for clothing segmentation trained on images and masks from the Viton dataset.
The document discusses gesture-based computing as an alternative to mouse input for human-computer interaction. It proposes a novel approach for implementing a real-time gesture recognition system capable of understanding commands based on analyzing the principal contour and fingertips of hand gestures. Vision-based gesture recognition techniques are discussed that do not require additional devices for users to interact with computers through natural hand motions.
SD-miner System to Retrieve Probabilistic Neighborhood Points in Spatial Dat...IOSR Journals
The document describes a proposed spatial data mining system called SD-Miner. SD-Miner consists of three main parts: a graphical user interface, an SD-Miner module for processing spatial data mining functions, and a data storage module. The SD-Miner module provides four spatial data mining functionalities: spatial clustering, spatial classification, spatial characterization, and spatio-temporal association rule mining. The document presents the architecture of SD-Miner and provides examples of using it to perform spatial clustering, classification, and characterization on spatial data from a database.
HUMAN COMPUTER INTERACTION ALGORITHM BASED ON SCENE SITUATION AWARENESScsandit
Implicit interaction based on context information is widely used and studied in the virtual scene.In context based human computer interaction, the meaning of action A is well defined. For instance, the right wave is defined turning paper or PPT in context B, And it mean volume up in context C. However, Select object in a virtual scene with multiple objects, context information is not fit. In view of this situation, this paper proposes using the least squares fitting curve beam to
predict the user's trajectory, so as to determine what object the user’s wants to operate .And fitting the starting position of the straight line according to the change of the discrete table. And
using the bounding box size control the Z variable to move in an appropriate location. Experimental results show that the proposed in this paper based on bounding box size to control
the Z variables get a good effect; by fitting the trajectory of a human hand, to predict the object that the subjects would like to operate. The correct rate is 88.6%.
Human Computer Interaction Algorithm Based on Scene Situation Awareness cscpconf
- The document proposes an algorithm for human-computer interaction based on scene situation awareness. It uses least squares fitting to predict the trajectory of a user's hand movement and determine what object the user wants to interact with.
- The algorithm first segments the user's hand from images and extracts features. It then fits the trajectory of hand movements with a nonlinear curve to predict the interaction target. Bounding box size is used to control movement in the z-axis for object selection.
- An experiment using this algorithm achieved an 88.6% correct rate in predicting the object a user intended to interact with based on fitting their hand trajectory.
Hand Gesture Recognition using OpenCV and Pythonijtsrd
Hand gesture recognition system has developed excessively in the recent years, reason being its ability to cooperate with machine successfully. Gestures are considered as the most natural way for communication among human and PCs in virtual framework. We often use hand gestures to convey something as it is non verbal communication which is free of expression. In our system, we used background subtraction to extract hand region. In this application, our PCs camera records a live video, from which a preview is taken with the assistance of its functionalities or activities. Surya Narayan Sharma | Dr. A Rengarajan "Hand Gesture Recognition using OpenCV and Python" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-2 , February 2021, URL: https://www.ijtsrd.com/papers/ijtsrd38413.pdf Paper Url: https://www.ijtsrd.com/computer-science/other/38413/hand-gesture-recognition-using-opencv-and-python/surya-narayan-sharma
A Survey Paper on Controlling Computer using Hand GesturesIRJET Journal
This document summarizes a survey paper on controlling computers using hand gestures. It discusses various techniques that have been used for hand gesture recognition in previous research papers. The paper reviews literature on hand gesture recognition methods based on sensor technology and computer vision. It describes applications of hand gesture recognition such as controlling media playback, scrolling web pages, and presenting slides. Common challenges with hand gesture recognition are also mentioned, such as dealing with complex backgrounds and lighting conditions. The goal of the paper is to perform a literature review on prominent techniques, applications, and difficulties in controlling computers using hand gestures.
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorQUESTJOURNAL
ABSTRACT:Tracking of moving objects that is called video tracking is used for measuring motion parameters and obtaining a visual record of the moving objects, it is an important area of application in image processing. In general there are two different approaches to obtain object tracking: the first is Recognition-based Tracking, and the second is the Motion-based Tracking. Video tracking system raises a wide possibility in today’s society. This system is used in various applications such as military, security, monitoring, robotic, and nowadays in dayto-day applications. However the video tracking systems still have many open problems and various research activities in a video tracking system are explores. This paper presents an algorithm for video tracking of any moving targets with the uses of contour based detection technique that depends on the sobel operator. The proposed system is suitable for indoor and outdoor applications. Our approach has the advantage of extending the applicability of tracking system and also, as presented here improves the performance of the tracker making feasible high frame rate video tracking. The goal of the tracking system is to analyze the video frames and estimate the position of a part of the input video frame (usually a moving object), our approach can detect, tracked any object more than one object and calculate the position of the moving objects. Therefore, the aim of this paper is to construct a motion tracking system for moving objects. Where, at the end of this paper, the detail outcome and result are discussed using experiments results of the proposed technique
Van der kamp.2011.gaze and voice controlled drawingmrgazer
This document describes a drawing application that is controlled using both gaze and voice inputs. The application allows users to draw various shapes like lines, rectangles, ellipses, and polygons using only their eyes to position the cursor and voice commands to activate the drawing. Previous gaze-based drawing tools required users to dwell their gaze at a location for a period of time to activate drawing, which caused delays and accidental activations. The proposed system aims to improve the user experience by removing the need to dwell gaze and only using gaze for positioning. The drawing application was implemented and evaluated through user trials. The results showed that while gaze and voice offered less control than traditional inputs, participants found it more enjoyable to use.
Hand Gesture Recognition System for Human-Computer Interaction with Web-Camijsrd.com
This paper represents a comparative study of exiting hand gesture recognition systems and gives the new approach for the gesture recognition which is easy cheaper and alternative of input devices like mouse with static and dynamic hand gestures, for interactive computer applications. Despite the increase in the attention of such systems there are still certain limitations in literature. Most applications require different constraints like having distinct lightning conditions, usage of a specific camera, making the user wear a multi-coloured glove or need lots of training data. The use of hand gestures provides an attractive alternative to cumbersome interface devices for human-computer interaction (HCI). This interface is simple enough to be run using an ordinary webcam and requires little training.
Gesture Based Interface Using Motion and Image Comparisonijait
This paper gives a new approach for movement of mouse and implementation of its functions using a real time camera. Here we propose to change the hardware design. Most of the existing technologies mainly depend on changing the mouse parts features like changing the position of tracking ball and adding more buttons. We use a camera, colored substance, image comparison technology and motion detection technology to control mouse movement and implement its functions (right click, left click, scrolling and double click) .
This document provides an overview of recent developments in object detection using AI robots. It explores several deep learning-based object detection techniques and their advantages over traditional computer vision methods. The paper discusses how object detection is used in robotics applications like grasping, manipulating, and navigation. It also presents the results of experiments conducted to evaluate an object detection system using a robot equipped with cameras and sensors. The system uses a Fast R-CNN algorithm combined with a Kalman filter for real-time object detection and tracking.
Similar to Bieg Eye And Pointer Coordination In Search And Selection Tasks (20)
Blignaut Visual Span And Other Parameters For The Generation Of HeatmapsKalle
Visual span and other parameters for the generation of heatmaps
1) Heatmaps are commonly used to visualize eye-tracking data but have disadvantages like difficulty interpreting colors and visual span is important but often overlooked.
2) An algorithm is proposed to generate heatmaps that allows the user to set the visual span parameter and customize transparency, color range, and probability of observing objects at different distances from fixations.
3) The algorithm assigns weights to fixations and pixels based on fixation duration, probability of observing pixels decreases with distance from fixation, and RGB color values are determined to represent weighted pixels on the heatmap.
Zhang Eye Movement As An Interaction Mechanism For Relevance Feedback In A Co...Kalle
This document discusses using eye movement data to infer image relevance in a content-based image retrieval system. An experiment was conducted where 10 participants viewed images from 101 categories to perform image searching tasks. Eye movement data including fixation duration, fixation count, and number of revisits was analyzed. The results show that fixation duration and count were significantly higher for positively relevant images compared to irrelevant images. This suggests eye movement data could provide an implicit form of relevance feedback to improve image retrieval systems.
A decision tree was then developed to predict user feedback during the tasks based on the eye tracking measures. It achieved over 87% accuracy, demonstrating the potential of using natural eye movement as a robust source of relevance feedback to bridge the semantic gap in
Yamamoto Development Of Eye Tracking Pen Display Based On Stereo Bright Pupil...Kalle
The intuitive user interfaces of PCs and PDAs, such as pen display and touch panel, have become widely used in recent times. In this study, we have developed an eye-tracking pen display based on the stereo bright pupil technique. First, the bright pupil camera was developed by examining the arrangement of cameras and LEDs for pen display. Next, the gaze estimation method was proposed for the stereo bright pupil camera, which enables one point calibration. Then, the prototype of the eyetracking pen display was developed. The accuracy of the system was approximately 0.7° on average, which is sufficient for human interaction support. We also developed an eye-tracking tabletop as an application of the proposed stereo bright pupil technique.
Wastlund What You See Is Where You Go Testing A Gaze Driven Power Wheelchair ...Kalle
Individuals with severe multiple disabilities have little or no opportunity to express their own wishes, make choices and move independently. Because of this, the objective of this work has been to develop a prototype for a gaze-driven device to manoeuvre powered wheelchairs or other moving platforms. The prototype has the same capabilities as a normal powered wheelchair, with two exceptions. Firstly, the prototype is controlled by eye movements instead of by a normal joystick. Secondly, the prototype is equipped with a sensor that stops all motion when the machine approaches an obstacle. The prototype has been evaluated in a preliminary clinical test with two users. Both users clearly communicated that they appreciated and had mastered the ability to control a powered wheelchair with their eye movements.
Vinnikov Contingency Evaluation Of Gaze Contingent Displays For Real Time Vis...Kalle
The visual field is the area of space that can be seen when an observer fixates a given point. Many visual capabilities vary with position in the visual field and many diseases result in changes in the visual field. With current technology, it is possible to build very complex real-time visual field simulations that employ gaze-contingent displays. Nevertheless, there are still no established techniques to evaluate such systems. We have developed a method to evaluate a system’s contingency by employing visual blind spot localization as well as foveal fixation. During the experiment, gaze-contingent and static conditions were compared. There was a strong correlation between predicted results and gaze-contingent trials. This evaluation method can also be used with patient populations and for the evaluation of gaze-contingent display systems, when there is need to evaluate a visual field outside of the foveal region.
Urbina Pies With Ey Es The Limits Of Hierarchical Pie Menus In Gaze ControlKalle
Pie menus offer several features which are advantageous especially for gaze control. Although the optimal number of slices per pie
and of depth layers has already been established for manual control, these values may differ in gaze control due to differences in spatial accuracy and congitive processing. Therefore, we investigated the layout limits for hierarchical pie menu in gaze control. Our user study indicates that providing six slices in multiple depth layers guarantees fast and accurate selections. Moreover, we compared two different methods of selecting a slice. Novices performed well with both, but selecting via selection borders produced better performance for experts than the standard dwell time selection.
Urbina Alternatives To Single Character Entry And Dwell Time Selection On Eye...Kalle
Eye typing could provide motor disabled people a reliable method of communication given that the text entry speed of current interfaces can be increased to allow for fluent communication. There are two reasons for the relatively slow text entry: dwell time selection requires waiting a certain time, and single character entry limits the maximum entry speed. We adopted a typing interface based on hierarchical pie menus, pEYEwrite [Urbina and Huckauf 2007] and included bigram text entry with one single pie iteration. Therefore, we introduced three different bigram building strategies.
Moreover, we combined dwell time selection with selection by borders, providing an alternative selection method and extra functionality. In a longitudinal study we compared participants performance during character-by-character text entry with bigram entry and with
text entry with bigrams derived by word prediction. Data showed large advantages of the new entry methods over single character text entry in speed and accuracy. Participants preferred selecting by
borders, which allowed them faster selections than the dwell time method.
Tien Measuring Situation Awareness Of Surgeons In Laparoscopic TrainingKalle
The study of surgeons’ eye movements is an innovative way of assessing skill and situation awareness, in that a comparison of eye movement strategies between expert surgeons and novices may show differences that can be used in training. Our preliminary study compared eye movements of 4 experts and
4 novices performing a simulated gall bladder removal task on a
dummy patient with an audible heartbeat and simulated vital signs displayed on a secondary monitor. We used a head-mounted Locarna PT-Mini eyetracker to record fixation locations during the operation. The results showed that novices concentrated so hard on the surgical
display that they were hardly able to look at the patient’s vital signs, even when heart rate audibly changed during the procedure. In comparison, experts glanced occasionally at the vitals monitor, thus being able to observe the patient condition.
Takemura Estimating 3 D Point Of Regard And Visualizing Gaze Trajectories Und...Kalle
The portability of an eye tracking system encourages us to develop a technique for estimating 3D point-of-regard. Unlike conventional methods, which estimate the position in the 2D image coordinates of the mounted camera, such a technique can represent richer gaze information of the human moving in the larger area. In this paper, we propose a method for estimating the 3D point-of-regard and a visualization technique of gaze trajectories under natural head movements for the head-mounted device. We employ visual SLAM technique to estimate head configuration and extract environmental information. Even in cases where the head moves dynamically, the proposed method could obtain 3D point-of-regard. Additionally, gaze trajectories are appropriately overlaid on the scene camera image.
Stevenson Eye Tracking With The Adaptive Optics Scanning Laser OphthalmoscopeKalle
Recent advances in high magnification retinal imaging have allowed for visualization of individual retinal photoreceptors, but these systems also suffer from distortions due to fixational eye motion. Algorithms developed to remove these distortions have the added benefit of providing arc second level resolution of the eye movements that produce them. The system also allows for visualization of targets on the retina, allowing for absolute retinal position measures to the level of individual cones. This paper will describe the process used to remove the eye movement artifacts and present analysis of their spectral characteristics. We find a roughly 1/f amplitude spectrum similar to that reported by Findlay (1971) with no evidence for a distinct
tremor component.
Stellmach Advanced Gaze Visualizations For Three Dimensional Virtual Environm...Kalle
Gaze visualizations represent an effective way for gaining fast insights into eye tracking data. Current approaches do not adequately support eye tracking studies for three-dimensional (3D) virtual environments. Hence, we propose a set of advanced gaze visualization techniques for supporting gaze behavior analysis in such environments. Similar to commonly used gaze visualizations for twodimensional
stimuli (e.g., images and websites), we contribute advanced 3D scan paths and 3D attentional maps. In addition, we introduce a models of interest timeline depicting viewed models, which can be used for displaying scan paths in a selected time segment. A prototype toolkit is also discussed which combines an implementation of our proposed techniques. Their potential for facilitating eye tracking studies in virtual environments was supported by a user study among eye tracking and visualization experts.
Skovsgaard Small Target Selection With Gaze AloneKalle
Accessing the smallest targets in mainstream interfaces using gaze
alone is difficult, but interface tools that effectively increase the size of selectable objects can help. In this paper, we propose a conceptual framework to organize existing tools and guide the development of new tools. We designed a discrete zoom tool and conducted a proof-of-concept experiment to test the potential of the framework and the tool. Our tool was as fast as and more accurate than the currently available two-step magnification tool. Our framework shows potential to guide the design, development, and testing of zoom tools to facilitate the accessibility of mainstream
interfaces for gaze users.
San Agustin Evaluation Of A Low Cost Open Source Gaze TrackerKalle
This document evaluates a low-cost open-source gaze tracking system using an off-the-shelf webcam. The performance of the gaze tracker was tested in an eye-typing task using two different typing applications. Participants were able to type between 3.56-6.78 words per minute depending on the application. A pilot study was also conducted with a user with severe motor impairments who was able to successfully type on a wall-projected interface using eye movements. The document describes the hardware and software components of the low-cost gaze tracking system and provides details on the experimental procedure and results of the eye-typing evaluation.
Ryan Match Moving For Area Based Analysis Of Eye Movements In Natural TasksKalle
Analysis of recordings made by a wearable eye tracker is complicated by video stream synchronization, pupil coordinate mapping, eye movement analysis, and tracking of dynamic Areas Of Interest (AOIs) within the scene. In this paper a semi-automatic system is developed to help automate these processes. Synchronization is accomplished
via side by side video playback control. A deformable eye template and calibration dot marker allow reliable initialization via simple drag and drop as well as a user-friendly way to correct the algorithm when it fails. Specifically, drift may be corrected by nudging the detected pupil center to the appropriate coordinates. In a case study, the impact of surrogate nature views on physiological health and perceived well-being is examined via analysis of gaze over images of nature. A match-moving methodology was developed to track AOIs for this particular application but is applicable toward similar future studies.
Rosengrant Gaze Scribing In Physics Problem SolvingKalle
Eye-tracking has been widely used for research purposes in fields such as linguistics and marketing. However, there are many possibilities of how eye-trackers could be used in other disciplines like physics. A part of physics education research deals with the differences between novices and experts, specifi-cally how each group solves problems. Though there has been a great deal of research about these differences there has been no research that focuses on noticing exactly where experts and no-vices look while solving the problems. Thus, to complement the past research, I have created a new technique called gaze scrib-ing. Subjects wear a head mounted eye-tracker while solving electrical circuit problems on a graphics monitor. I monitor both scan patterns of the subjects and combine that with videotapes of their work while solving the problems. This new technique has yielded new information and elaborated on previous studies.
Qvarfordt Understanding The Benefits Of Gaze Enhanced Visual SearchKalle
In certain applications such as radiology and imagery analysis, it is important to minimize errors. In this paper we evaluate a structured inspection method that uses eye tracking information as a feedback mechanism to the image inspector. Our two-phase method starts with a free viewing phase during which gaze data is collected. During the next phase, we either segment the image, mask previously seen areas of the image, or combine the two techniques, and repeat the search. We compare the different methods
proposed for the second search phase by evaluating the inspection method using true positive and false negative rates, and subjective workload. Results show that gaze-blocked configurations reduced the subjective workload, and that gaze-blocking without segmentation showed the largest increase in true positive identifications and the largest decrease in false negative identifications of previously unseen objects.
Prats Interpretation Of Geometric Shapes An Eye Movement StudyKalle
This paper describes a study that seeks to explore the correlation between eye movements and the interpretation of geometric shapes. This study is intended to inform the development of an eye tracking interface for computational tools to support and enhance the natural interaction required in creative design. A common criticism of computational design tools is that they do not enable manipulation of designed shapes according to all perceived features. Instead the manipulations afforded are limited by formal structures of shapes. This research examines the potential for eye movement data to be used to recognise and make available for manipulation the perceived features in shapes. The objective of this study was to analyse eye movement data with the intention of recognising moments in which an interpretation of shape is made. Results suggest that fixation duration and saccade amplitude prove to be consistent indicators of shape interpretation.
Porta Ce Cursor A Contextual Eye Cursor For General Pointing In Windows Envir...Kalle
Eye gaze interaction for disabled people is often dealt with by designing ad-hoc interfaces, in which the big size of their elements compensates for both the inaccuracy of eye trackers and the instability of the human eye. Unless solutions for reliable eye cursor control are employed, gaze pointing in ordinary graphical operating environments is a very difficult task. In this paper we present an eye-driven cursor for MS Windows which behaves differently according to the “context”. When the user’s gaze is perceived within the desktop or a folder, the cursor can be discretely shifted from one icon to another. Within an application window or where there are no icons, on the contrary, the cursor can be continuously and precisely moved. Shifts in the four directions (up, down, left, right) occur through dedicated buttons. To increase user awareness of the currently pointed spot on the screen while continuously moving the cursor, a replica of the spot is provided within the active direction button, resulting in improved pointing performance.
Pontillo Semanti Code Using Content Similarity And Database Driven Matching T...Kalle
Laboratory eyetrackers, constrained to a fixed display and static (or accurately tracked) observer, facilitate automated analysis of fixation data. Development of wearable eyetrackers has extended environments and tasks that can be studied at the expense of automated analysis. Wearable eyetrackers provide 2D point-of-regard (POR) in scene-camera coordinates, but the researcher is typically interested in some high-level semantic property (e.g., object identity, region, or material) surrounding individual fixation points. The synthesis of POR into fixations and semantic information remains a labor-intensive manual task, limiting the application of wearable eyetracking.
We describe a system that segments POR videos into fixations and allows users to train a database-driven, object-recognition system. A correctly trained library results in a very accurate and semi-automated translation of raw POR data into a sequence of objects, regions or materials.
Park Quantification Of Aesthetic Viewing Using Eye Tracking Technology The In...Kalle
The purpose of this study is to explore how the viewers’ previous training is related to their aesthetic viewing in various interactions with the form and the context, in relation to apparel design. Berlyne’s two types of exploratory behavior, diversive and specific, provided a theoretical framework to this study. Twenty female subjects (mean age=21, SD=1.089) participated. Twenty model images, posed by a male and a female model, were shown on an eye-tracker screen for 10 seconds each. The findings of this study verified Berlyne’s concepts of visual exploration. One of the different findings from Berlyne’s theory was that the untrained viewers’ visual attention tended to be more significantly focused on peripheral areas of visual interest, compared to the trained viewers, while there was no significant difference on the central, foremost areas of visual interest between the two groups. The overall aesthetic viewing patterns were also identified.
2. screen was shown to inform participants of the current target’s trial took considerably longer (mean: 5.5 s, std. dev.: 0.60, 95%
color and shape. For reasons of consistency, this was also done CI of difference: 1.5-2.3 s, t(10)=10.5, p<0.01). Fewer fixations
in task T1. This preparatory screen contained an instance of the were made during the single-target task T1 (mean: 6.0, std. dev.:
target item and an “ok” button which had to be clicked in order 0.78) compared to the search task T2 (mean: 14.4, std. dev.:
to start the trial. For tasks T1 and T2, the button and item in- 2.37, 95% CI of difference: 6.7-10.0, t(10)=11.1, p<0.01). Thus,
stance were located at a constant position to the right of the it took participants approximately 2 seconds longer to complete
screen. The button was at the exact same position as the “done” task T2. This time was likely spent on the search for the target
button in the trial display. The item instance was above the item and eye movements required for item identification.
button. For task T3, item and button were shown in the center of
the screen. With this arrangement, gaze and pointer position Since almost 40% of trial completion time seems to be spent on
were at an approximate constant starting position for each trial search, how do users coordinate pointing movements during this
(near the “done” button on the right side for T1 and T2 or center time? Analysis of pointer movement in task T2 shows that the
of the screen for T3). pointer was already close to the target when the first fixation on
the target item was made. For instance, in more than 80% of
trials, the pointer had already moved half of its initial distance to
the target (Figure 2 line a). For the simple selection task T1, the
pointer’s distance to the target was generally larger. Still, in
almost 60% of trials, the pointer had already moved more than
half of the initial distance (Figure 2 line b). In task T2, the me-
dian latency between the first fixation on the target item and the
beginning of the pointer movement was -1.7 s. In T1, the median
latency was -190 ms. In more than 60% of trials of T2 the poin-
ter movement was initiated almost 1 s before the target had been
fixated (Figure 3 line a). In trials of T1, the latency was shorter
but still mostly negative.
Figure 1 Schematic representation of the experimental tasks.
2.3 Apparatus
Participants used an optical computer mouse to complete the
task (Logitech Optical Wireless). The experimental computer
ran Ubuntu Linux version 8.10. Participants sat 100 cm in front
of a large back-projected screen which displayed the graphical
stimuli at a rate of 120 Hz (resolution 1400 × 1050 px, size
94.1° × 89.1° visual angle, 2,15 × 1,97 m). While carrying out
the tasks, participants’ gaze was tracked with a mobile eye-
tracking system. The system allowed for unconstrained head
motion by combining a head mounted eye-tracker (SR Research
Eyelink II) with an infrared motion capturing system (Vicon MX
13 with four cameras) using the libGaze tracking library [Her-
holz et al. 2008]. With this system, gaze was computed in terms
of screen coordinates at 120 Hz.
Figure 2 Cumulative histogram of the distance (as proportions
2.4 Procedure of the initial pointer-to-target distance of approximately 800 px,
68° visual angle) between the pointer and the target item at the
Each experimental session lasted around 90 minutes with short time of the first fixation on the target.
pauses every 15 minutes and regular recalibration of the eye-
tracker before each block. Prior to data collection, the tasks were Two explanations for these findings seem probable. First, partic-
explained, demonstrated to the participants, and one round of ipants moved the pointer immediately once the trial started by
practice trials was presented. Each participant then carried out using knowledge from previous trials about the approximate
five blocks of tasks. Each block started with 10 trials of task T1, target location (T1) or grid location (T2). This movement re-
followed by 10 trials of T2, and 3 trials of task T3. quired no visual guidance and was carried out while the eye
made a saccade to the target to guide the final targeting phase of
3 Results the movement. Similar findings were also reported for hand
movements by van Donkelaar et al. [2000].
It is well known from experiments on visual search that search
A second explanation could be that, in the case of T2, partici-
for combinations of features (e.g. shape and color) take more
pants parallelized search and pointer movement. To investigate
time as a function of the number of items and difficulty of dis-
this, we analyzed pointer movements during task T3. In this task
crimination. In addition, the number of fixations is correlated
target locations could not be predicted from memory. A typical
with search difficulty [Zelinsky and Sheinberg 1997]. We com-
example of eye scanpaths and pointer motion is shown in Figure
pared trial completion times and number of fixations of tasks T1
5. Trajectories for the vertical and horizontal dimension along
and T2 to see whether an effect of the added requirement to
the time axis are shown in Figure 5 b and c. From the inspection
search an item prior to selection was present. The mean trial
of individual trial data it seems that the pointer roughly followed
completion time for a simple drag-and-drop trial (T1) was 3.5 s
the scanpath throughout the trial. To test this hypothesis, we
(std. dev.: 0.45). In the search condition (T2) completing the
90
3. calculated the mean Euclidean distance between the location of p<0.01). This indicates that participants generally moved the
the mouse pointer and eye gaze across all trials of T3 for all pointer while searching and did not adopt a turn-taking strategy.
participants. The mean Euclidean distance was minimal when
the mouse pointer signal was shifted a few hundred milliseconds
forward in time (see Figure 4). Even though the extent of the
shift varied between participants (see participant means in Fig-
ure 4) it was positive for all participants. A positive shift can be
conceived as a shift of the mouse pointer path to the left in
Figure 5 b and c. This means the mouse pointer generally lagged
behind the eye movements but it does not give any clues to
whether it did so consistently throughout the trial.
Figure 3 Cumulative histogram of the latency between a fixation
on the target and movement initiation (the point in time when the
pointer was moved more than 40 px from the starting position).
Figure 5 Typical eye scanpath and path of pointer motion for a
single trial of task T3. Item locations are shown as gray circles
(color/shape coding not shown) and target items are circles
colored in red. Selection actions are depicted as blue stars.
4 Conclusion
In this study, we investigated the coordination of eye move-
ments and mouse pointer movements during a range of search
and selection tasks. In performing these tasks, participants use
two main eye-hand coordination strategies. First, when the
approximate location of the target is known, participants per-
form pointer movements that are initiated without eye guidance,
possibly in parallel to an eye movement to the target item.
Figure 4 Mean distance of the pointer to the gaze point on the Second, in tasks requiring visual search for a target item prior to
screen across trials of task T3 as a function of the amount of selection, users parallelize search and pointer movements. This
phase shift of the mouse pointer “signal”. Pictograms indicate behavior might have two functions: First, in order to utilize
effect of shift. “slack motor time” while looking for the target, users roughly
follow their scanpath to keep the pointer close to the searched
To analyze whether the pointer followed gaze throughout a trial, area. Thereby they minimize the amplitude of the acquisition
we implemented a model of pointer motion in which the pointer movement, once a target is found. Second, users might employ
was moved directly from selection point to selection point and the pointer as an active reference tool for marking potential
remained motionless during the periods in between. This model target items or as a reference for keeping track of the search
simulates the turn-taking strategy assumed by some assistive area. Such “deictic” pointing behavior [Ballard et al. 1997] has
interaction techniques in which a user moves the mouse to ac- also been observed in menu selection and web browsing tasks
quire a target but keeps it stationary while searching. Compared [Cox and Silva, 2006; Rodden et al. 2008]. Further analyses and
against actual pointer motion the mean Euclidean distance experiments are required to clarify the function of the observed
across all trials and participants was significantly larger for this behavior in the current task.
turn-taking model (95% CI difference: 238-289 px, t(10)=23.0,
91
4. Our results also show that the eye reaches the targeted item prior COX, A. L., AND SILVA, M. M. 2006. The role of mouse move-
to the pointer. This is in line with the “target gaze” behavior ments in interactive search. In Proceedings of the 28th Annual
identified by previous studies [Smith et al. 2000; Law et al. Meeting of the Cognitive Science Society. Lawrence Erlbaum
2004]. However, our findings also imply that the eyes fixate on Associates, 1156-1161.
the target rather late in the pointing process if the approximate
target location is known. This gives rise to the assumption that DREWES, H., AND SCHMIDT, A. 2009. The MAGIC Touch: Com-
eye fixations are only needed for precise targeting motions bining MAGIC-Pointing with a Touch-Sensitive Mouse. In
during the final pointing phase. We did not find evidence for the Human-Computer Interaction – INTERACT 2009 (Proceed-
“following” or “switching” behaviors described by Smith and ings of 12th IFIP TC 13 International Conference). Springer,
colleagues [2000]. We cannot exclude the possibility that these 415-428.
behaviors may occur. In the context of search tasks, however,
our data suggest that participants consistently follow their gaze FLEETWOOD, M. D., AND BYRNE, M. D. 2008. Modeling the
with the mouse pointer rather than the other way round. visual search of displays: a revised ACT-R model of icon
search based on eye-tracking data. Human-Computer Interac-
These findings also have implications for the design of opti- tion, 21, 2, 153–197.
mized pointing techniques, for instance, gaze-assisted tech-
niques that reduce the amplitude of the pointing movement [e.g. HERHOLZ, S., CHUANG, L. L., TANNER, T. G., BÜLTHOFF, H. H.,
Zhai et al. 1999]. An implicit assumption of these techniques is AND FLEMING, R.W. 2008. LibGaze Real-time gaze-tracking
that the eyes precede a pointer movement so that movement time of freely moving observers for wall-sized displays. In Pro-
savings could be gained by moving the pointer to the point of ceedings of the Vision, Modeling, and Visualization Work-
gaze. Results of the present study indicate that users tend to shop (VMV). IOS Press, 101-110.
minimize pointing amplitudes with little effort by carrying out
approximate pointer motions in parallel to visual search when LAW, B., ATKINS, S. M., KIRKPATRICK, A. E., LOMAX, A. J., AND
the target is unknown or initiating movements even before fixat- MACKENZIE, C. L. 2004. Eye Gaze Patterns Differentiate No-
ing the target when the target location is known. Both scenarios vice and Experts in a Virtual Laparoscopic Surgery Training
are very common in many user interface tasks. Locations of Environment. In Proceedings of the 2004 symposium on eye
interface elements such as window controls are often well tracking research & applications. ACM, 41-47.
known whereas content elements such as links on a web page
must often be searched [Rodden et al. 2008]. In fact it is hard to MACKENZIE, S. I. 1992. Fitts' Law as a Research and Design
construct a scenario where no knowledge is available or no Tool in Human-Computer Interaction. Human-Computer In-
search is needed. It might therefore be reasonable to explore teraction, 7, 1, 91-139.
other optimization techniques, for instance, techniques that do
not minimize movement amplitudes but the needed endpoint MCGUFFIN, M., AND BALAKRISHNAN, R. 2002. Acquisition of
precision, for instance using target-expansion [e.g. McGuffin expanding targets. In Proceedings of CHI, ACM Press, 57-64.
and Balakrishnan 2002].
NEGGERS, S. F. W., AND BEKKERING, H. 2000. Ocular Gaze is
5 Acknowledgements Anchored to the Target of an Ongoing Pointing Movement.
Journal of Neurophysiology, 83, 639-651.
This project is supported by the Baden-Württemberg Informa-
RÄIHÄ, K.-J., AND ŠPAKOV, O. 2009. Disambiguating Ninja
tion Technology program (BW-FIT) through the state of Baden-
Cursors with Eye Gaze. In Proceedings of CHI. ACM, 1411-
Württemberg, Germany and the World Class University pro-
1414.
gram through the National Research Foundation of Korea
funded by the Ministry of Education, Science and Technology
RODDEN, K., FU, X., AULA, A., AND SPIRO, I. 2008. Eye-Mouse
(R31-2008-000-10008-0).
Coordination Patterns on Web Search Results Pages. In
Proceedings of CHI extended abstracts. ACM, 2997-3002.
References
SMITH, B. A., HO, J., ARK, W., AND ZHAI, S. 2000. Hand Eye
BALAKRISHNAN, R. 2004. “Beating” Fitts’ law: virtual enhance- Coordination Patterns in Target Selection. In Proceedings of
ments for pointing facilitation. International Journal of Hu- the 2000 symposium on eye tracking research & applications.
man-Computer Studies, 61, 857-874. ACM, 117-122.
BALLARD, D. H., HAYHOE, M. M., POOK, P. K., AND RAO, R. P. N. VAN DONKELAAR, P., AND STAUB, J. 2000. Eye-hand coordina-
1997. Deictic codes for the embodiment of cognition. Beha- tion to visual versus remembered targets. Experimental Brain
vioral and Brain Sciences, 20, 723-767. Research, 133, 414-418.
BIEG, H., CHUANG, L. L., AND REITERER, H. 2009. Gaze-Assisted ZELINSKY, G. J., AND SHEINBERG, D. L. 1997. Eye Movements
Pointing for Wall-Sized Displays. In Human-Computer Inte- During Parallel-Serial Visual Search. Journal of Experimental
raction – INTERACT 2009 (Proceedings of 12th IFIP TC 13 Psychology: Human Perception and Performance, 23, 1, 244-
International Conference). Springer, 9-12. 262.
BLANCH, R., AND ORTEGA, M. 2009. Rake cursor: improving ZHAI, S., MORIMOTO, C., IHDE, S. 1999. Manual and Gaze Input
pointing performance with concurrent input channels. In Pro- Cascaded (MAGIC) Pointing. Proceedings of CHI. ACM,
ceedings of CHI. ACM, 1415-1418. 246 – 253.
92