The document describes a laser pointer interaction system for manipulating 3D medical images on a large display in operating rooms. The system uses a webcam to detect the laser spot and track its movements. It recognizes two types of gestures - a circle gesture with two directions and a line gesture with four directions. The system was tested on 15 subjects performing rotation and zooming tasks on a 3D model. Experiments showed that the dynamic time warping algorithm recognized gestures with 89.6% accuracy and was faster than the 1$ recognizer algorithm. The laser pointer system provides a potential hands-free alternative to traditional mouse/keyboard control in operating rooms.
PEDESTRIAN DETECTION IN LOW RESOLUTION VIDEOS USING A MULTI-FRAME HOG-BASED D...AM Publications
Detecting pedestrians in low resolution videos is a challenging task, due to the small size of pedestrians in the images and the limited information. In practical outdoor surveillance scenarios the pedestrian size is usually small. Existing state-of-the-art pedestrian detection methods that use histogram of oriented gradient (HOG) features have poor performance in this problem domain. To compensate for the lack of information in a single frame, we propose a novel detection method that recognizes pedestrians in a short sequence of frames. Namely, we take the single-frame HOG-based detector and extend it to multiple frames. Our detector is applied to regions containing potential moving objects. In the case of video taken from a moving camera on an aerial platform, video stabilization is first performed to register the frames. A classifier is then applied to features extracted from spatio-temporal volumes surrounding the potential moving objects. On challenging stationary and aerial video datasets, our detection accuracy outperforms several state-of-the-art algorithms.
Joint Roughness and Wrinkle Detection Using Gabor Filtering and Dynamic Line ...CSCJournals
Skin quality analysis is a continuously increasing field for the detection of skin dryness, wrinkles and other skin conditions for the recommendation of skin-care products. Each method reveals valuable information on age, disease and general habits. Contributions are being made from companies, but many methods make use of extensive image data and diagnostic equipment that are too expensive to obtain. These would involve people visiting skin-care shops for a diagnosis, which is often time consuming and troublesome. A joint technique is proposed to locate skin roughness as part of an optimized wrinkle detection technique based on the capabilities of mobile devices without the use of diagnostic tools. The technique makes use of thresholding, shape analysis and dynamic line tracking for the analysis of skin roughness and wrinkles.
FINGERPRINT CLASSIFICATION MODEL BASED ON NEW COMBINATION OF PARTICLE SWARM O...IAEME Publication
Fingerprint is one of the broadly utilized biometric distinguishing proof to
recognize the persons due steady and worthiness. The classification of fingerprints
gives indexing to the dataset to decrease the seeking and matching manner. There
researches have been developed numerous approaches for fingerprint classification
model, for example, the Neural Network (NN) methods and support vector machine.
Nevertheless, there are a lot of developments to develop the classification procedure.
In this paper, a new classification model based on PSO-SVM technique is used to
achieve fingerprint classification process. The proposed methodology comprises three
phases: preprocessing phases, feature extraction phase and classification phase using
the proposed PSO-SVM model. The CASIA V5 fingerprint dataset and a property
dataset are adopted in this study to test the proposed model. The experimental results
showed that the proposed PSO-SVM classification model was better than the standard
SVM in terms of accuracy, sensitivity and specificity for both datasets.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Review of face detection systems based artificial neural networks algorithmsijma
Face detection is one of the most relevant applications of image processing and biometric systems.
Artificial neural networks (ANN) have been used in the field of image processing and pattern recognition.
There is lack of literature surveys which give overview about the studies and researches related to the using
of ANN in face detection. Therefore, this research includes a general review of face detection studies and
systems which based on different ANN approaches and algorithms. The strengths and limitations of these
literature studies and systems were included also.
In recent days, skin cancer is seen as one of the most Hazardous form of the cancer found in Humans. Skin Cancer is a malignant tumor that grows in the skin cells. It can be affected mostly by the reason of skin burn caused by sunlight. Early detection and treatment of Skin cancer can significantly improve patient outcome. Automatic detection is one of the most challenging research areas that can be used for early detection of such vital cancer. A person’s in which they have inadequate amount of melanoma will be exposed to the risk of sun burns and the ultra violet rays from the sun will be affected that body. Malignant melanomas is a type of melanoma that has irregular borders, color variations so analyze the shape, color and texture of the skin lesion is important for the early detection. It can have the components of an automated image analysis module, which contains image acquisition, hair detection and exclusion, lesion segmentation, feature extraction and finally classification. Finally the result show that the system is efficient achieving classification of the lesion as either melanoma or Non melanoma causes.
PEDESTRIAN DETECTION IN LOW RESOLUTION VIDEOS USING A MULTI-FRAME HOG-BASED D...AM Publications
Detecting pedestrians in low resolution videos is a challenging task, due to the small size of pedestrians in the images and the limited information. In practical outdoor surveillance scenarios the pedestrian size is usually small. Existing state-of-the-art pedestrian detection methods that use histogram of oriented gradient (HOG) features have poor performance in this problem domain. To compensate for the lack of information in a single frame, we propose a novel detection method that recognizes pedestrians in a short sequence of frames. Namely, we take the single-frame HOG-based detector and extend it to multiple frames. Our detector is applied to regions containing potential moving objects. In the case of video taken from a moving camera on an aerial platform, video stabilization is first performed to register the frames. A classifier is then applied to features extracted from spatio-temporal volumes surrounding the potential moving objects. On challenging stationary and aerial video datasets, our detection accuracy outperforms several state-of-the-art algorithms.
Joint Roughness and Wrinkle Detection Using Gabor Filtering and Dynamic Line ...CSCJournals
Skin quality analysis is a continuously increasing field for the detection of skin dryness, wrinkles and other skin conditions for the recommendation of skin-care products. Each method reveals valuable information on age, disease and general habits. Contributions are being made from companies, but many methods make use of extensive image data and diagnostic equipment that are too expensive to obtain. These would involve people visiting skin-care shops for a diagnosis, which is often time consuming and troublesome. A joint technique is proposed to locate skin roughness as part of an optimized wrinkle detection technique based on the capabilities of mobile devices without the use of diagnostic tools. The technique makes use of thresholding, shape analysis and dynamic line tracking for the analysis of skin roughness and wrinkles.
FINGERPRINT CLASSIFICATION MODEL BASED ON NEW COMBINATION OF PARTICLE SWARM O...IAEME Publication
Fingerprint is one of the broadly utilized biometric distinguishing proof to
recognize the persons due steady and worthiness. The classification of fingerprints
gives indexing to the dataset to decrease the seeking and matching manner. There
researches have been developed numerous approaches for fingerprint classification
model, for example, the Neural Network (NN) methods and support vector machine.
Nevertheless, there are a lot of developments to develop the classification procedure.
In this paper, a new classification model based on PSO-SVM technique is used to
achieve fingerprint classification process. The proposed methodology comprises three
phases: preprocessing phases, feature extraction phase and classification phase using
the proposed PSO-SVM model. The CASIA V5 fingerprint dataset and a property
dataset are adopted in this study to test the proposed model. The experimental results
showed that the proposed PSO-SVM classification model was better than the standard
SVM in terms of accuracy, sensitivity and specificity for both datasets.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Review of face detection systems based artificial neural networks algorithmsijma
Face detection is one of the most relevant applications of image processing and biometric systems.
Artificial neural networks (ANN) have been used in the field of image processing and pattern recognition.
There is lack of literature surveys which give overview about the studies and researches related to the using
of ANN in face detection. Therefore, this research includes a general review of face detection studies and
systems which based on different ANN approaches and algorithms. The strengths and limitations of these
literature studies and systems were included also.
In recent days, skin cancer is seen as one of the most Hazardous form of the cancer found in Humans. Skin Cancer is a malignant tumor that grows in the skin cells. It can be affected mostly by the reason of skin burn caused by sunlight. Early detection and treatment of Skin cancer can significantly improve patient outcome. Automatic detection is one of the most challenging research areas that can be used for early detection of such vital cancer. A person’s in which they have inadequate amount of melanoma will be exposed to the risk of sun burns and the ultra violet rays from the sun will be affected that body. Malignant melanomas is a type of melanoma that has irregular borders, color variations so analyze the shape, color and texture of the skin lesion is important for the early detection. It can have the components of an automated image analysis module, which contains image acquisition, hair detection and exclusion, lesion segmentation, feature extraction and finally classification. Finally the result show that the system is efficient achieving classification of the lesion as either melanoma or Non melanoma causes.
Skin Cancer Detection using Image Processing in Real Timeijtsrd
Machine learning is a fascinating topic its astonishing how a small change in the evaluation values may result in an unfathomable number of outcomes. The goal of this study is to develop a model that uses image processing to identify skin cancer. We will later use the model in real life through an android application. Sunami Dasgupta | Soham Das | Sayani Hazra Pal "Skin Cancer Detection using Image-Processing in Real-Time" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-6 , October 2021, URL: https://www.ijtsrd.com/papers/ijtsrd46384.pdf Paper URL : https://www.ijtsrd.com/computer-science/artificial-intelligence/46384/skin-cancer-detection-using-imageprocessing-in-realtime/sunami-dasgupta
ABSTRACT : A hardware/software (HW/SW) co-design approach developing a contact lenses try-on system via the use of the Field Programmable Gate Array (FPGA) [1] is proposed in this paper. To achieve an interactive user-friendly environment, the menu of contact lenses and the result that shows how it looks when wearing the contact lens are displayed on the LCD touch panel directly. After capturing the user’s image via CMOS sensors, image processing is employed to determine user’s irises from the image. A particle filter algorithm is then introduced to track the irises so that the contact lenses chosen by the user can be overlapped on the image to provide a natural examination of the effect of the contact lenses try-on system. Experimental results show good performances of the proposed system. Keywords: FPGA, Eye Tracking, Particle Filter, Nios II, Image Processing
A Survey : Iris Based Recognition SystemsEditor IJMTER
The security is one of the important aspect of today's life. Iris recognization is one of the leading
research of security which is used to identify the individual person. Usually iris based biometric is more better
than other biometric in terms of accuracy, fast, stability, uniqueness. The iris recognition system works by
capturing and storing biometric information and then compare scanned copy of iris biometric with the stored iris
images in the database. There are several Iris Based Recognition Systems are developed so far. In this paper we
presented several iris techniques and create a base for our future roadmap.
A Proposed Framework for Robust Face Identification SystemAhmed Gad
Presentation of IEEE paper "A Proposed Framework for Robust Face Identification System"
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?tp=&arnumber=7030929&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel7%2F7022265%2F7030901%2F07030929.pdf%3Farnumber%3D7030929
Find me on:
AFCIT
http://www.afcit.xyz
YouTube
https://www.youtube.com/channel/UCuewOYbBXH5gwhfOrQOZOdw
Google Plus
https://plus.google.com/u/0/+AhmedGadIT
SlideShare
https://www.slideshare.net/AhmedGadFCIT
LinkedIn
https://www.linkedin.com/in/ahmedfgad/
ResearchGate
https://www.researchgate.net/profile/Ahmed_Gad13
Academia
https://www.academia.edu/
Google Scholar
https://scholar.google.com.eg/citations?user=r07tjocAAAAJ&hl=en
Mendelay
https://www.mendeley.com/profiles/ahmed-gad12/
ORCID
https://orcid.org/0000-0003-1978-8574
StackOverFlow
http://stackoverflow.com/users/5426539/ahmed-gad
Twitter
https://twitter.com/ahmedfgad
Facebook
https://www.facebook.com/ahmed.f.gadd
Pinterest
https://www.pinterest.com/ahmedfgad/
M.Sc. Thesis - Automatic People Counting in Crowded ScenesAhmed Gad
This thesis proposes a real-time automatic people crowd density estimation method for overcoming the non-linearity problem, working with different densities and scales, and enhancing the prediction error. To cover most of the properties of the crowded scene, a newly used combination of features is proposed that includes segmented region properties, texture, edge, and SIFT keypoints. Edge strength is a suggested for use.
Identification and Rejection of Defective Ceramic Tile using Image Processing...IJMTST Journal
Manual Ceramic Tile inspection process is tedious if human operator is employed to look for defective tiles and their elimination. The plain ceramic tiles often have the following types of defects viz Cracks, Blobs and pin holes [1]. The fatigue of human operator deteriorates the quality of the tile being produced. In this paper a novel and simple automatic tile defect identification and elimination system is proposed. The proposed system is built around MATLAB and ARDUINO. The systems performance is evaluated in terms of accuracy and time taken for detection. The Proposed system promises superior performance when compared to the other existing system
Brain Tumor Area Calculation in CT-scan image using Morphological Operationsiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Eye tracking system has played a significant role in many of today’s applications ranging from military
applications to automotive industries and healthcare sectors. In this paper, a novel system for eye tracking and
estimation of its direction of movement is performed. The proposed system is implemented in real time using an
arduino uno microcontroller and a zigbee wireless device. Experimental results show a successful eye tracking and
movement estimation in real time scenario using the proposed hardware interface.
Skin cancer is a term given to the uncontrolled
growth of strange skin cells. It occurs whenever unrepaired
DNA damages to skin cells trigger mutations, or any other
genetic defects, that lead the skin cells to multiply readily
and form malignant tumors. Image processing is a
commonly used method for skin cancer detection from the
appearance of the affected area on the skin. The input to the
system is that the skin lesion image so by applying novel
image process techniques, it analyses it to conclude about
the presence of skin cancer. The Lesion Image analysis tools
checks for the various Melanoma parameters Like
Asymmetry, Border, Colour, Diameter, (ABCD rule), etc. by
texture, size and form analysis for image segmentation and
have stages. The extracted feature parameters are
accustomed classify the image as traditional skin and
malignant melanoma cancerlesion.
Artificial Neural Network (ANN) is one of the
important branches of Artificial Intelligence, which has
been accepted as a brand-new technology in computer
science for image processing. Neural Networks is currently
the area of interest in medicine, particularly in the fields of
radiology, urology, cardiology, oncology, etc. Neural
Network plays a vital role in an exceedingly call network. It
has been used to analyze Melanoma parameters Like
Asymmetry, Border, Colour, Diameter, etc. which are
calculated using MATLAB from skin cancer images
intending to developing diagnostic algorithms that might
improve triage practices in the emergency department.
Using the ABCD rules for melanoma skin cancer, we use
ANN in the classification stage. Initially, we train the
network with known target values. The network is well
trained with 96.9% accuracy, and then the unknown values
are tested for the cancer classification. This classification
method proves to be more efficient for skin cancer
classification
Analysis of Butterworth and Chebyshev Filters for ECG Denoising Using WaveletsIOSR Journals
Abstract: A wide area of research has been done in the field of noise removal in Electrocardiogram signals.. Electrocardiograms (ECG) play an important role in diagnosis process and providing information regarding heart diseases. In this paper, we propose a new method for removing the baseline wander interferences, based on discrete wavelet transform and Butterworth/Chebyshev filtering. The ECG data is taken from non-invasive fetal electrocardiogram database, while noise signal is generated and added to the original signal using instructions in MATLAB environment. Our proposed method is a hybrid technique, which combines Daubechies wavelet decomposition and different thresholding techniques with Butterworth or Chebyshev filter. DWT has good ability to decompose the signal and wavelet thresholding is good in removing noise from decomposed signal. Filtering is done for improved denoising performence. Here quantitative study of result evaluation has been done between Butterworth and Chebyshev filters based on minimum mean squared error (MSE), higher values of signal to interference ratio and peak signal to noise ratio in MATLAB environment using wavelet and signal processing toolbox. The results proved that the denoised signal using Butterworth filter has a better balance between smoothness and accuracy than the Chebvshev filter. Keywords: Electrocardiogram, Discrete Wavelet transform, Baseline Wandering, Thresholding, Butterworth, Chebyshev
Skin Cancer Detection using Image Processing in Real Timeijtsrd
Machine learning is a fascinating topic its astonishing how a small change in the evaluation values may result in an unfathomable number of outcomes. The goal of this study is to develop a model that uses image processing to identify skin cancer. We will later use the model in real life through an android application. Sunami Dasgupta | Soham Das | Sayani Hazra Pal "Skin Cancer Detection using Image-Processing in Real-Time" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-6 , October 2021, URL: https://www.ijtsrd.com/papers/ijtsrd46384.pdf Paper URL : https://www.ijtsrd.com/computer-science/artificial-intelligence/46384/skin-cancer-detection-using-imageprocessing-in-realtime/sunami-dasgupta
ABSTRACT : A hardware/software (HW/SW) co-design approach developing a contact lenses try-on system via the use of the Field Programmable Gate Array (FPGA) [1] is proposed in this paper. To achieve an interactive user-friendly environment, the menu of contact lenses and the result that shows how it looks when wearing the contact lens are displayed on the LCD touch panel directly. After capturing the user’s image via CMOS sensors, image processing is employed to determine user’s irises from the image. A particle filter algorithm is then introduced to track the irises so that the contact lenses chosen by the user can be overlapped on the image to provide a natural examination of the effect of the contact lenses try-on system. Experimental results show good performances of the proposed system. Keywords: FPGA, Eye Tracking, Particle Filter, Nios II, Image Processing
A Survey : Iris Based Recognition SystemsEditor IJMTER
The security is one of the important aspect of today's life. Iris recognization is one of the leading
research of security which is used to identify the individual person. Usually iris based biometric is more better
than other biometric in terms of accuracy, fast, stability, uniqueness. The iris recognition system works by
capturing and storing biometric information and then compare scanned copy of iris biometric with the stored iris
images in the database. There are several Iris Based Recognition Systems are developed so far. In this paper we
presented several iris techniques and create a base for our future roadmap.
A Proposed Framework for Robust Face Identification SystemAhmed Gad
Presentation of IEEE paper "A Proposed Framework for Robust Face Identification System"
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?tp=&arnumber=7030929&url=http%3A%2F%2Fieeexplore.ieee.org%2Fiel7%2F7022265%2F7030901%2F07030929.pdf%3Farnumber%3D7030929
Find me on:
AFCIT
http://www.afcit.xyz
YouTube
https://www.youtube.com/channel/UCuewOYbBXH5gwhfOrQOZOdw
Google Plus
https://plus.google.com/u/0/+AhmedGadIT
SlideShare
https://www.slideshare.net/AhmedGadFCIT
LinkedIn
https://www.linkedin.com/in/ahmedfgad/
ResearchGate
https://www.researchgate.net/profile/Ahmed_Gad13
Academia
https://www.academia.edu/
Google Scholar
https://scholar.google.com.eg/citations?user=r07tjocAAAAJ&hl=en
Mendelay
https://www.mendeley.com/profiles/ahmed-gad12/
ORCID
https://orcid.org/0000-0003-1978-8574
StackOverFlow
http://stackoverflow.com/users/5426539/ahmed-gad
Twitter
https://twitter.com/ahmedfgad
Facebook
https://www.facebook.com/ahmed.f.gadd
Pinterest
https://www.pinterest.com/ahmedfgad/
M.Sc. Thesis - Automatic People Counting in Crowded ScenesAhmed Gad
This thesis proposes a real-time automatic people crowd density estimation method for overcoming the non-linearity problem, working with different densities and scales, and enhancing the prediction error. To cover most of the properties of the crowded scene, a newly used combination of features is proposed that includes segmented region properties, texture, edge, and SIFT keypoints. Edge strength is a suggested for use.
Identification and Rejection of Defective Ceramic Tile using Image Processing...IJMTST Journal
Manual Ceramic Tile inspection process is tedious if human operator is employed to look for defective tiles and their elimination. The plain ceramic tiles often have the following types of defects viz Cracks, Blobs and pin holes [1]. The fatigue of human operator deteriorates the quality of the tile being produced. In this paper a novel and simple automatic tile defect identification and elimination system is proposed. The proposed system is built around MATLAB and ARDUINO. The systems performance is evaluated in terms of accuracy and time taken for detection. The Proposed system promises superior performance when compared to the other existing system
Brain Tumor Area Calculation in CT-scan image using Morphological Operationsiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Eye tracking system has played a significant role in many of today’s applications ranging from military
applications to automotive industries and healthcare sectors. In this paper, a novel system for eye tracking and
estimation of its direction of movement is performed. The proposed system is implemented in real time using an
arduino uno microcontroller and a zigbee wireless device. Experimental results show a successful eye tracking and
movement estimation in real time scenario using the proposed hardware interface.
Skin cancer is a term given to the uncontrolled
growth of strange skin cells. It occurs whenever unrepaired
DNA damages to skin cells trigger mutations, or any other
genetic defects, that lead the skin cells to multiply readily
and form malignant tumors. Image processing is a
commonly used method for skin cancer detection from the
appearance of the affected area on the skin. The input to the
system is that the skin lesion image so by applying novel
image process techniques, it analyses it to conclude about
the presence of skin cancer. The Lesion Image analysis tools
checks for the various Melanoma parameters Like
Asymmetry, Border, Colour, Diameter, (ABCD rule), etc. by
texture, size and form analysis for image segmentation and
have stages. The extracted feature parameters are
accustomed classify the image as traditional skin and
malignant melanoma cancerlesion.
Artificial Neural Network (ANN) is one of the
important branches of Artificial Intelligence, which has
been accepted as a brand-new technology in computer
science for image processing. Neural Networks is currently
the area of interest in medicine, particularly in the fields of
radiology, urology, cardiology, oncology, etc. Neural
Network plays a vital role in an exceedingly call network. It
has been used to analyze Melanoma parameters Like
Asymmetry, Border, Colour, Diameter, etc. which are
calculated using MATLAB from skin cancer images
intending to developing diagnostic algorithms that might
improve triage practices in the emergency department.
Using the ABCD rules for melanoma skin cancer, we use
ANN in the classification stage. Initially, we train the
network with known target values. The network is well
trained with 96.9% accuracy, and then the unknown values
are tested for the cancer classification. This classification
method proves to be more efficient for skin cancer
classification
Analysis of Butterworth and Chebyshev Filters for ECG Denoising Using WaveletsIOSR Journals
Abstract: A wide area of research has been done in the field of noise removal in Electrocardiogram signals.. Electrocardiograms (ECG) play an important role in diagnosis process and providing information regarding heart diseases. In this paper, we propose a new method for removing the baseline wander interferences, based on discrete wavelet transform and Butterworth/Chebyshev filtering. The ECG data is taken from non-invasive fetal electrocardiogram database, while noise signal is generated and added to the original signal using instructions in MATLAB environment. Our proposed method is a hybrid technique, which combines Daubechies wavelet decomposition and different thresholding techniques with Butterworth or Chebyshev filter. DWT has good ability to decompose the signal and wavelet thresholding is good in removing noise from decomposed signal. Filtering is done for improved denoising performence. Here quantitative study of result evaluation has been done between Butterworth and Chebyshev filters based on minimum mean squared error (MSE), higher values of signal to interference ratio and peak signal to noise ratio in MATLAB environment using wavelet and signal processing toolbox. The results proved that the denoised signal using Butterworth filter has a better balance between smoothness and accuracy than the Chebvshev filter. Keywords: Electrocardiogram, Discrete Wavelet transform, Baseline Wandering, Thresholding, Butterworth, Chebyshev
Abstract: An audio mixer amplifier is a device that translates a signal of one frequency band to another. It will accept many inputs at different frequencies and generate an output of the combination or sum of the frequencies. The mixer circuit provides good gain to weak audio signals. It can be used in front of an R.F. oscillator to make an R.F. receiver that is very sensitive to sound. Each input can be independently controlled by a variable resistor. There is also a provision for a balance control to fade out signal while simultaneously fading in the other. Key Words: Audio Mixer, Frequency, Signal, Circuit
Implementation of Efficiency CORDIC Algorithmfor Sine & Cosine GenerationIOSR Journals
Abstract: This paper presents an area-time efficient coordinate rotation digital computer (CORDIC) algorithm that completely eliminates the scale-factor. A generalized micro-rotation selection technique based on high speed most-significant-1-detection obviates the complex search algorithms for identifying the micro-rotations. This algorithm is redefined as the elementary angles for reducing the number of CORDIC iterations. Compared to the existing re-cursive architectures the proposed one has 17% lower slice-delay product on Xilinx Spartan XC2S200E device. The CORDIC processor pro-vides the flexibility to manipulate the number of iterations depending on the accuracy, area and latency requirements. Index Terms—coordinate rotation digital computer (CORDIC), cosine/sine, field-programmable gate array(FPGA),most-significant-1, recursive architecture.
IOSR Journal of Electrical and Electronics Engineering(IOSR-JEEE) is an open access international journal that provides rapid publication (within a month) of articles in all areas of electrical and electronics engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electrical and electronics engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
IOSR Journal of Mathematics(IOSR-JM) is an open access international journal that provides rapid publication (within a month) of articles in all areas of mathemetics and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in mathematics. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Nasal Parameters of Ibibio and Yakurr Ethnic Groups of South South NigeriaIOSR Journals
The study involved 400 subjects, 200 each of the two ethnic groups(100 males and 100 females). The age range was 18-35years. Subjects with facial deformities or surgeries that involved the nasal region were excluded from the study. Subjects that have both parents and grandparents of each ethnic group and who have lived in that environment for at least the first 18years of their lifetime were selected for the study. The nasal length and width were measured using spreading calipers and nasal indices deducted from these measurements. Test for significance was done using the students’ t-test. Results showed the nasal indices of the Ibibio males and females to be 86.58±1.20 and 81.75±1.14 respectively and for the Yakurr males and females, it was 77.76±0.82 and 102.27±1.13 respectively. There was significant ethnic and gender differences in all the nasal parameters (nasal length, nasal width and nasal indices) at p<0.05. From the nasal indices the nose type of the Ibibio males is platyrrhine while that of the Ibibio females is mesorrhine, the Yakurr males also have mesorrhine nose type while the Yakurr females have platyrrhine nose type. With this result, nasal parameters could be a useful tool in gender and ethnic differentiation between the Ibibio and Yakurr ethnic groups.
Chemical Reaction Effects on Free Convective Flow of a Polar Fluid from a Ver...IOSR Journals
This article deals with a study of two dimensional free convective flow of a polar fluid through a porous medium due to combined effects of thermal and mass diffusion in presence of a chemical reaction of first order. The objective of the present investigation is to analyze the free convective flow in the presence of prescribed wall heat flux and mass flux condition. The governing partial differential equations are non-dimensionalized and transformed into a system of non-similar equations. The resulting coupled nonlinear partial differential equations are solved numerically under appropriate transformed boundary conditions using an implicit finite difference scheme in combination with quasilinearisation technique. Computations are performed for a wide range of values of the various governing flow parameters of the velocity, angular velocity, temperature and species concentration profiles and results are presented graphically. The numerical results for local skin friction coefficient, couple stress coefficient, local Nusselt number and local Sherwood number are also presented. The obtained results are compared with previously published work and were to be in excellent agreement. The study reveals that the flow characteristics are profoundly influenced by the polar effects
Speech Recognized Automation System Using Speaker Identification through Wire...IOSR Journals
Abstract : This paper discusses the methodology for a project named “Speech Recognized Automation System using Speaker Identification through wireless communication”. This project gives the design of Automation system using wireless communication and speaker recognition using Matlab code. Straightforward programming interface of Matlab makes it an ideal tool for speech analysis in project. This automation system is useful for home appliances as well as in industry. This paper discusses the overall design of a wireless automation system which is built and implemented. The speech recognition centers on recognition of speech commands stored in data base of Matlab and it is matched with incoming voice command of speaker. Mel Frequency Cepstral Coefficient (MFCC) algorithm is used to recognize the speech of speaker and to extract features of speech. It uses low-power RF ZigBee transceiver wireless communication modules which are relatively cheap. This automation system is intended to control lights, fans and other electrical appliances in a home or office using speech commands like Light, Fan etc. Further, if security is not big issue then Speech processor is used to control the appliances without speaker identification. Keywords — Automation system, MATLAB code, MFCC, speaker identification, ZigBee transceiver.
Hand Gesture Recognition System for Human-Computer Interaction with Web-Camijsrd.com
This paper represents a comparative study of exiting hand gesture recognition systems and gives the new approach for the gesture recognition which is easy cheaper and alternative of input devices like mouse with static and dynamic hand gestures, for interactive computer applications. Despite the increase in the attention of such systems there are still certain limitations in literature. Most applications require different constraints like having distinct lightning conditions, usage of a specific camera, making the user wear a multi-coloured glove or need lots of training data. The use of hand gestures provides an attractive alternative to cumbersome interface devices for human-computer interaction (HCI). This interface is simple enough to be run using an ordinary webcam and requires little training.
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorQUESTJOURNAL
ABSTRACT:Tracking of moving objects that is called video tracking is used for measuring motion parameters and obtaining a visual record of the moving objects, it is an important area of application in image processing. In general there are two different approaches to obtain object tracking: the first is Recognition-based Tracking, and the second is the Motion-based Tracking. Video tracking system raises a wide possibility in today’s society. This system is used in various applications such as military, security, monitoring, robotic, and nowadays in dayto-day applications. However the video tracking systems still have many open problems and various research activities in a video tracking system are explores. This paper presents an algorithm for video tracking of any moving targets with the uses of contour based detection technique that depends on the sobel operator. The proposed system is suitable for indoor and outdoor applications. Our approach has the advantage of extending the applicability of tracking system and also, as presented here improves the performance of the tracker making feasible high frame rate video tracking. The goal of the tracking system is to analyze the video frames and estimate the position of a part of the input video frame (usually a moving object), our approach can detect, tracked any object more than one object and calculate the position of the moving objects. Therefore, the aim of this paper is to construct a motion tracking system for moving objects. Where, at the end of this paper, the detail outcome and result are discussed using experiments results of the proposed technique
Autonomous Camera Movement for Robotic-Assisted Surgery: A SurveyIJAEMSJORNAL
In the past decade, Robotic-Assisted Surgery (RAS) has become a widely accepted technique as an alternative to traditional open surgery procedures. The best robotic assistant system should combine both human and robot capabilities under the human control. As a matter of fact robot should collaborate with surgeons in a natural and autonomous way, thus requiring less of the surgeons’ attention. In this survey, we provide a comprehensive and structured review of the robotic-assisted surgery and autonomous camera movement for RAS operation. We also discuss several topics, including but not limited to task and gesture recognition, that are closely related to robotic-assisted surgery automation and illustrate several successful applications in various real-world application domains. We hope that this paper will provide a more thorough understanding of the recent advances in camera automation in RSA and offer some future research directions.
Analytical framework for optimized feature extraction for upgrading occupancy...IJECEIAES
The adoption of the occupancy sensors has become an inevitable in commercial and non-commercial security devices, owing to their proficiency in the energy management. It has been found that the usages of conventional sensors is shrouded with operational problems, hence the use of the Doppler radar offers better mitigation of such problems. However, the usage of Doppler radar towards occupancy sensing in existing system is found to be very much in infancy stage. Moreover, the performance of monitoring using Doppler radar is yet to be improved more. Therefore, this paper introduces a simplified framework for enriching the event sensing performance by efficient selection of minimal robust attributes using Doppler radar. Adoption of analytical methodology has been carried out to find that different machine learning approaches could be further used for improving the accuracy performance for the feature that has been extracted in the proposed system of occuancy system.
Multi touch interactive screen, MIE CompetitionHadeel M. Yusef
Multi touch Interactive Screen is a Graduation Project for the 2010 year,computer science, Ain Shams University, It is extended project of Interactive Wall 2009. Multi touch interactive Screen won the first rank in Made in Egypt Competition, IEEE Egypt section
Human-machine interactions based on hand gesture recognition using deep learn...IJECEIAES
Human interaction with computers and other machines is becoming an increasingly important and relevant topic in the modern world. Hand gesture recognition technology is an innovative approach to managing computers and electronic devices that allows users to interact with technology through gestures and hand movements. This article presents deep learning methods that allow you to efficiently process and classify hand gestures and hand gesture recognition technologies for interacting with computers. This paper discusses modern deep learning methods such as convolutional neural networks (CNN) and recurrent neural networks (RNN), which show excellent results in gesture recognition tasks. Next, the development and implementation of a human-machine interaction system based on hand gesture recognition is discussed. System architectures are described, as well as technical and practical aspects of their application. In conclusion, the article summarizes the research results and outlines the prospects for the development of hand gesture recognition technology to improve human- machine interaction. The advantages and limitations of the technology are analyzed, as well as possible areas of its application in the future.
Real time hand gesture recognition system for dynamic applicationsijujournal
Virtual environments have always been considered as a means for more visceral and efficient human computer interaction by a diversified range of applications. The spectrum of applications includes analysis of complex scientific data, medical training, military simulation, phobia therapy and virtual prototyping.
Evolution of ubiquitous computing, current user interaction approaches with keyboard, mouse and pen are
not sufficient for the still widening spectrum of Human computer interaction. Gloves and sensor based trackers are unwieldy, constraining and uncomfortable to use. Due to the limitation of these devices the useable command set based diligences is also limited. Direct use of hands as an input device is an
innovative method for providing natural Human Computer Interaction which has its inheritance from textbased interfaces through 2D graphical-based interfaces, multimedia-supported interfaces, to full-fledged multi-participant Virtual Environment (VE) systems. Conceiving a future era of human-computer
interaction with the implementations of 3D application where the user may be able to move and rotate objects simply by moving and rotating his hand - all without help of any input device.
Real time hand gesture recognition system for dynamic applicationsijujournal
Virtual environments have always been considered as a means for more visceral and efficient human computer interaction by a diversified range of applications. The spectrum of applications includes analysis of complex scientific data, medical training, military simulation, phobia therapy and virtual prototyping. Evolution of ubiquitous computing, current user interaction approaches with keyboard, mouse and pen are not sufficient for the still widening spectrum of Human computer interaction. Gloves and sensor based trackers are unwieldy, constraining and uncomfortable to use. Due to the limitation of these devices the useable command set based diligences is also limited. Direct use of hands as an input device is an innovative method for providing natural Human Computer Interaction which has its inheritance from textbased interfaces through 2D graphical-based interfaces, multimedia supported interfaces, to full-fledged multi-participant Virtual Environment (VE) systems. Conceiving a future era of human-computer interaction with the implementations of 3D application where the user may be able to move and rotate objects simply by moving and rotating his hand - all without help of any input device. The research effort centralizes on the efforts of implementing an application that employs computer vision algorithms and gesture recognition techniques which in turn results in developing a low cost interface device for interacting with objects in virtual environment using hand gestures. The prototype architecture of the application comprises of a central computational module that applies the camshift technique for tracking of hands and its gestures. Haar like technique has been utilized as a classifier that is creditworthy for locating hand position and classifying gesture. The patterning of gestures has been done for recognition by mapping the number of defects that is formed in the hand with the assigned gestures. The virtual objects are produced using Open GL library. This hand gesture recognition technique aims to substitute the use of mouse for interaction with the virtual objects. This will be useful to promote controlling applications like virtual games, browsing images etc in virtual environment using hand gestures.
Robot operating system based autonomous navigation platform with human robot ...TELKOMNIKA JOURNAL
In emerging technologies, indoor service robots are playing a vital role for people who are physically challenged and visually impaired. The service robots are efficient and beneficial for people to overcome the challenges faced during their regular chores. This paper proposes the implementation of autonomous navigation platforms with human-robot interaction which can be used in service robots to avoid the difficulties faced in daily activities. We used the robot operating system (ROS) framework for the implementation of algorithms used in auto navigation, speech processing and recognition, and object detection and recognition. A suitable robot model was designed and tested in the Gazebo environment to evaluate the algorithms. The confusion matrix that was created from 125 different cases points to the decent correctness of the model.
Social Service Robot using Gesture recognition techniqueChristo Ananth
A robot is a machine that can automatically do a task or a series of tasks based on its programming and environment. They are artificially built machines or devices that can perform activities with utmost accuracy and precision minimizing time constraints. Service robots are technologically advanced machines deployed to service and maintain certain activities. Research findings convey the essential fact that serving robots are now being deployed worldwide. Social robotics is one such field that heavily involves an interaction between humans and an artificially built machine. These man-built machines interact with humans and can also understand social terms and words. Modernization has bought changes in design and mechanisms due to this ever-lasting growth in technology and innovation. Therefore, food industries are also dynamically adapting to the new changes in the field of automation to reduce human workload and increase the quality of service. Deployment of a robot in the food industries which help to aid deaf and mute people who face social constraints is an evergrowing challenge faced by engineers for the last few decades. Moreover, a contactless form of speedy service system which accomplishes its task with at most precision and reduced complexity is a feat yet to be perfected. Preservation of personal hygiene, a better quality of service, and reduced labour costs is achieved.
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Generating a custom Ruby SDK for your web service or Rails API using Smithyg2nightmarescribd
Have you ever wanted a Ruby client API to communicate with your web service? Smithy is a protocol-agnostic language for defining services and SDKs. Smithy Ruby is an implementation of Smithy that generates a Ruby SDK using a Smithy model. In this talk, we will explore Smithy and Smithy Ruby to learn how to generate custom feature-rich SDKs that can communicate with any web service, such as a Rails JSON API.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
D017511825
1. IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 5, Ver. I (Sep. – Oct. 2015), PP 18-25
www.iosrjournals.org
DOI: 10.9790/0661-17511825 www.iosrjournals.org 18 | Page
Laser pointer interaction for 3D angiography operation
Dina Ahmed**
, Ayman Atia*
, Essam A. Rashed**
, and
Mostafa-Samy M. Mostafa*
*
HCI-LAB, Department of CS, Faculty of Computers and Information, Helwan
University, Helwan, Egypt
**Image Science-LAB, Department of Mathematics, Faculty of Science, Suez Canal University, Ismailia, Egypt
Abstract: A primary challenge for creating an interactive display in operating room (OR) is in the definition of
control methods that are efficient and easy to learn for the physician. Apart from traditional input methods such
as mouse and keyboard, we present a system that utilizes the laser pointer to manipulate directly the 3D medical
images of angiography on LCD display. This work is based on laser detection, tracking the spot, and laser
gesture recognition using an ordinary laser pointer. The laser spot is captured by a web camera and its location
recognized with the help of the computer vision techniques. The system includes two moving operations to
execute commands: ”circle gesture” with two directions and ”line gesture” with four directions. The recorded
laser gestures are then recognized using two different algorithms: dynamic time wrapping (DTW) and one
dollar (1$) recognizer. Our experiments show that the DTW algorithm performs better with an overall accuracy
of %89.6 and is faster than the 1$ algorithm. This paper describes the proposed system, its implementation, and
an experiment that is performed to evaluate it.
Keywords:Direct interaction, laser pointers, 3D image manipulation, human-computer interaction.
I. Introduction
Large displays are increasingly being critical and prevalent in our daily activities. Previous studies have
shown their benefits in several tasks such as navigation [1], multitasking [2]. As well as they encourage
communication and collaboration among members of a group in a shared area [3]. User interaction with large
displays is one of the important topics in human-computer interaction (HCI) research community [4]. Recently,
many applications of a large interactive display have been explored. It had a significant impact on different
fields such as medical image processing, advertising, information visualization, public collaboration, and
education classrooms [5, 6]. As large display becomes widespread and ubiquitous, it presents some challenges
that can not be easily handled by using traditional input devices such as mouse and keyboard [7]. However,
users of large displays are often standing, working up close to the screen and may want to walk unrestrictedly in
front of it. It is awkward and not practical to ask them to hold the mouse while manipulating the objects on the
screen.
Typically, any person in a meeting room, where an object is shown on a large display, uses a laser
pointer device to point to some details displayed on the screen. It is useful to identify ways in which improving
the use of laser pointers. For instance, employing laser pointer as an interactive tool to do some operations such
as open a new application, zooming, and panning images. This idea could be achieved if the laser pointer is
somehow used as a device that would be able to go like the mouse on the screen. Several methods have been
proposed to interact with large displays using laser pointers. For example, circling gesture [8] and dwell time
[9], these methods cover differentfields such as Windows operations [10]. However, it is still difficult to use the
laser pointer as an efficient input device because of problems such as the hand jitter problem, detection errors,
and latency.
In this paper, we attempt to use the laser pointer in a medical application that can be useful in operating
rooms. Hepatic angiography is a study of an X-ray of the blood vessels that supply the liver. The procedure uses
a thin and flexible catheter that is placed into a blood vessel through a small cut. A trained doctor called an
Interventional radiologist usually performs the procedure. The proposed work is part of a project entitled
IAngio, this project is a research activity that aims to develop a three-dimensional (3D) imaging system of
hepatic angiography.
In the OR, surgeons can not use devices like mice and keyboards. In such situation, the surgeon usually
asks another staff member to help him to interact with the display (e.g. Zoom in/out images). The latency
between surgeon decisions and actions to be performed by the other member motivated our work. As a
substitute, laser pointer technique is proposed in this paper for this particular problem. An interactive system is
presented to help surgeons in the OR to control the reconstructed 3D image on the LCD using laser pointer
gestures, which can be executed at a distance from the screen.
2. Laser pointer interaction for 3D angiography operation
DOI: 10.9790/0661-17511825 www.iosrjournals.org 19 | Page
The system only uses an inexpensive laser pointer with an on/off button and a web camera to capture
the movement of the laser spot on the screen. The main contribution here is to develop an algorithm that has the
ability to understand the laser pointer gestures as commands to interact with the objects shown on the screen
depending on the presence and the absence of the laser spot, and explore different application that can be used
with those gestures.
The rest of the paper is arranged as follows: Section 2 summarizes the related work, Section 3 describes
the system overview of our research and illustrates the proposed approach. Section 4 and Section 5 present the
experiment and the results.
II. Related Work
Interaction with large displays falls mainly under two categories: remote and up-close interaction. The
classic problem with up-close interactions is that users are required to walk up to the screen to interact with
objects that are within arms reach. Anastasia et al. [11] proposed the Vacuum system, as a potential solution to
the problem of access to remote areas of the display that are difficult to reach. It brings remote objects closer to
the user for viewing and manipulation. However, when users step back from the display to view the contents of
the entire screen, they can no longer interact with the objects until they step forward to touch the screen.
One of the solutions is to use remote interactive techniques that allow users to interact with large
displays from a distance. Instead of learning completely different new methods to interact using new scenarios,
it would be better to adopt the natural channels of communication that are familiar to people in their daily life,
e.g. gesture, eye-gaze, speech, and physical tools [12, 13, 14, 15]. One way explored is through the use of laser
pointer [16] as an input device. As it allows direct interactions, and the users can perform all the operations by
small wrist motion [17, 18]. However, all of these systems adopted a compatible mouse interface. Where, the
user holds the laser pointer with a fixed posture for a certain amount of time on a specific spot on the display to
click anything. This technique is called the ”holding technique”.
Myers [19] showed that it is difficult for a user to keep a laser pointer fixed on a specific area of a
screen for a long time as the beam is unsteady, and users cannot turn the device on or off at will. Shizuki et al.
[20] attempted to move away from this method where the proposed system depends on the crossing technique
where the users execute commands through using the four peripheral screen areas. Hisamatsu et al. [21] also
combined the crossing technique with another one that’s based on moving gesture, i.e. the user can draw circles
by laser pointer to select objects. Bi et al. [22] used laser pointers with buttons as interaction devices for one or
more displays. Such systems are more suitable for multi-user interactions. However, the additional electronics
requirements reduce the advantage of laser pointers: low cost and ubiquity.
One of the important issues in the laser pointer system is its ability to detect the laser spot. Threshold
method is considered one of the popular ways in the detection process. It attempts to extract the object of
interest by finding the brightest spot in the image [23]. However, Shin et al. [24] claimed that: 1) high contrast
level may cause the laser spot to appear as a glow, which prevents the correct detection. 2) A change in the
environment lighting may cause the algorithm to misinterpret the spot position, and 3) laser spot may not be the
brightest spot on the display due to lighting effects or the existence of other bright colors. Sugawara and Miki
[25] used a special laser pointer called WDM with a visible laser beam and an invisible infrared laser beam.
Then, they attached a band pass filter to the camera so that only infrared wavelength can pass through. A hybrid
technique [26] may also be used to solve the problem of the threshold method, but the algorithm may be too
slow for a real-time application.
Background subtraction is another simple and popular method for laser spot detection [27]. One of the
disadvantages of this method is that any foreground object remains static for a period will disappear as the
algorithm will misunderstand it as a background object.
In this paper, two methods are evaluated to select the one that is well suited to our system for the laser
detection process: thresholding and background subtraction. The proposed system is based on several steps:
laser spot detection, tracking laser spot, and laser gesture recognition that will be described in detail in the
following section.
III. System Overview
The proposed system is composed of a large screen, a laser pointer and a laptop PC connected to a web
camera to detect the laser pointer position (see Figure 1(a)). For the camera, the system requires a low-cost web
camera yet adequate for initial tests, it can deliver up to 30 frames per second. The camera is used to capture the
image of the screen and then in this captured image, the system detects the laser pointer and identifies its
corresponding position on the screen. The image is shown over the large display (see Figure 1(c)), and the
surgeon controls the reconstructed 3D image through performing laser gestures for zooming, rotating, and
flipping (see Figure 1(b)).
3. Laser pointer interaction for 3D angiography operation
DOI: 10.9790/0661-17511825 www.iosrjournals.org 20 | Page
Figure 1: Laser pointer control system.
3.1 System Details
In this section, the proposed system details are discussed that endow computers with the ability to
understand the context in which laser gestures are made. After the image has been captured by the web camera,
it is considered as the input of the laser detection algorithm. Laser spot detection step is performed to identify
which one of the detected foregrounds is the actual laser spot. First, thresholding process is performed to
separate the image into a foreground object and background. Every frame is converted from RGB to HSV color
space, where the color of interest (i.e. red or green color) is filtered between a minimum and maximum
threshold values as seen in Figure 2. However, the threshold approach failed to detect the laser spot with other
bright colors (see Figure 3(a)). Second, the background subtraction method is performed where a frame is
compared to another frame in which a significant difference between those frames is identified as the
foreground object. This method achieves good results in distinguishing the laser spot as illustrated in Figure
3(b). After testing the two approaches, the background subtraction is perfectly suited to the proposed system
based on the assumption that the background of our environment is entirely static.
Figure 2: Laser spot detection.
Second, some morphological operations are performed such as ”Erode” and ”Dilate” to remove noise
and improve the appearance of the laser spot. After the algorithm detects the presence of the laser spot on the
image, it returns its position corresponding to the screen grid. The moments method used to calculate the
position of the center of the laser spot. The first order spatial moments calculated around x-axis and y-axis and
the zero order central moments of the binary image. Where, the zero order central moments of the binary image
are equal to the white area of the image in pixels (i.e. the noise of the binary image also should be at the
minimum level to get accurate results).
4. Laser pointer interaction for 3D angiography operation
DOI: 10.9790/0661-17511825 www.iosrjournals.org 21 | Page
For high-quality results, there are few constraints in the system to be setup. The proposed system uses the
following basic assumptions:
1. The web camera is supposed to be fixed and motionless.
2. The brightness and exposure settings of the camera are reduced in order to improve the detectability of . the
laser spot
Figure 3: The two approaches with other bright colors.
One of the challenges that is considered to be a part of the laser gesture recognition system is about
identifying the start and the end points of the intended gesture. As the user moves from one laser gesture to
another, he performs several unintended movements with his hand that links the two consecutive intended
gestures. The system may attempt to recognize this inescapable intermediate move as a meaningful one. To
solve this problem, each gesture is recognized with the laser-on and laser-of events. The start point is recognized
when the laser button is pressed. In contrast, the end point is recognized when the user releases the laser pointer
button. However, the user may attempt to perform laser on/off events only to point to some details on the screen,
and he/she does not mean to interact with the large display.
To differentiate between intended and unintended gestures and based on our user analysis empirical
study, we assume that every intended laser gesture takes about 1-2 seconds to be performed. Thus, any gesture
exceeds that time is considered as an unintended gesture.
Finally, to recognize the laser gestures, two different algorithms are tested in order to choose the best
algorithm for the proposed system. The first algorithm is called dynamic time warping (DTW) algorithm. The
DTW is a time series alignment algorithm developed basically for speech recognition. It aligns two sequences of
feature vectors by warping the time axis iteratively until an optimal match between the two sequences is found.
The second one is called 1$ recognizer algorithm. The 1$ uni-stroke recognizer is a 2-D single-stroke recognizer
designed for rapid prototyping of gesture-based user interfaces. It is an instance-based nearest-neighbor
classifier with the Euclidean scoring function, i.e., a geometric template matcher.
In the recognition process, the system receives a continuous stream of coordinates, the trajectory of
each intended laser gesture is recorded. The data that is collected from the web camera is used to recognize
gestures such as the circle in the two rotational directions and the line with the four directions: up, down, left,
and right. Every gesture will then be compared with a set of predefined training data, or prototypes (in the form
of templates) to recognize which gesture is being signed. A label corresponding to the recognized gesture will
be displayed on the large display.
IV. Experiment
In order to understand how laser pointer performs in interventional procedures, the experimental
evaluation has been performed. The main research questions are:
1. Whether the laser pointer system is a potential substitute for the conventional input system in ORs.
2. Whether the laser pointer system can facilitate the surgeons in dealing with the 3D models that is shown on
the large display.
Thus, our goal of this experiment was to measure the total time, including error correction, that is spent
on executing the interactions with the large display using the laser pointer compared to other techniques that are
executed in the OR in a conventional way, i.e. the surgeon asks another staff member to help him to interact
with the display using mouse and keyboard. Second, measuring the accuracy of recognizing proposed gestures
5. Laser pointer interaction for 3D angiography operation
DOI: 10.9790/0661-17511825 www.iosrjournals.org 22 | Page
for controlling the 3D model. The classification accuracy is evaluated in the experiment between two
algorithms: DTW and 1$ recognizer.
4.1 Experiment Setup
The laboratory setup composed of a USB camera with 30 fps and resolution of 640X480. A green laser
pointer (wavelength 532 nm, output power about 500 mW) is used, and the data is collected on a laptop PC
running Microsoft Windows 7 with 2.4 GHz processor and 8 GB RAM. The software was written in OpenCV
C++. The experiment was performed by 15 subjects, there were 5 males and 10 females aged between 19 and
25. All participants were frequent users of mouse-and-windows based systems, but none of them had any
previous experience with interactions through laser pointers. The users were asked to stand close to the large
display at a distance of 1.5 m (see Figure 5).
Each user is asked to use three input ways: the traditional way, i.e. input with a mouse and keyboard,
and the laser pointer per each classifier (DTW and 1$ recognizer) in order to complete two tasks as shown in
Figure 4:
Task 1: Rotating the 3D model over X and Y axis using four gestures: right-line, left-line, up-line, and down-
line.
Task 2: Zooming in/out the 3D model using two gestures: circle-clockwise and circle-counterclockwise.
Figure 4: Laser command mappings.
Users have a single training session before performing the experiment to familiarize themselves with
the laser pointer technique. Each user was briefed on the goals of the experiment, and the system was
demonstrated by the experimenter. Then, each of them should complete 4 sessions per each technique. Each
session consisted of all six commands and has a different order for the performing gestures.
Figure 5: Laboratory experiment using the traditional method and the laser pointer.
V. Results and Discussion
Experimental data shows that mouse-based interactions are faster than the laser pointer either with the
DTW or the 1$ (Figure 6). However, the average time cost to complete the total task with the mouse and the
keyboard is just two seconds faster than that with a laser pointer using the DTW recognizer. Considering that,
the latency problem or the slowness of the spot recognition can be solved by using better and faster cameras
connections, we are satisfied with the performance of the laser pointer with DTW recognizer relative to the
traditional method.
6. Laser pointer interaction for 3D angiography operation
DOI: 10.9790/0661-17511825 www.iosrjournals.org 23 | Page
Figure 6: Total time cost for the three input ways including error correction.
The averaging accuracy of DTW is %89.6, it means that the system misinterprets the laser pointer
gestures at a rate of %10. Generally, it is noticed that there are some obvious sources of error in the DTW
system. For instance, some circle gestures and line gestures (i.e. usually left-line and right-line) are misclassified
as down-line or up-line. The expected reason is that some users perform lines with a slightly skewness down or
up. Similarly, circles with a very small diameter also misunderstood by the system. In the experiment of the
DTW, we observed the hand jitter problem but, it was not a major problem as reported in existing literature [9].
It did not have a significant effect in our experiment because users were not asked to hold the laser steady for a
long time. That is, drawing circles or lines is relatively easier than keeping a laser spot fixed on a specified
location for a period of time.
Figure 7: Time cost in seconds.
As shown in Figures 7 and 8, the laser pointer with the 1$ recognizer including error correction is
almost twice as slow as the mouse and the averaging accuracy of 1$ is %75. The algorithm performs well and
achieves good results in differentiating between the line shape and the circle shape, but it also has some
limitations. The 1$ can not distinguish gestures whose identities depend on aspect ratios, locations, or specific
orientations. This means that separating up-lines from down-lines or left-lines from right-lines is not possible
without modifying the conventional algorithm. This explains why the 1$ comes last in the total time results.
When the users make mistakes, they need extra time to correct their errors as we give them two trial for error
correction.
Figure 9 displays the number of errors for all participants in each session for both algorithms. It is
observed that the errors in DTW in session 3 and 4 are much less than those in session 1 and 2 which indicates
that a user can be expected to improve in the usage of the laser pointer with the DTW. However, we can not
expect that in the 1$ classifier. There was no discernible difference in user performance between any of the four
sessions. After calculating the number of errors for all participants in each session, the largest number occurs
during session 2.
7. Laser pointer interaction for 3D angiography operation
DOI: 10.9790/0661-17511825 www.iosrjournals.org 24 | Page
The obtained data from participants gestures are then analyzed using an ANOVA test. First, a one-way
ANOVA test is performed in regards to the time of the total task for the three input ways. The test indicates that
there is a significant difference between the three algorithms (p<0.05). However, the ANOVA analysis only
indicates that if there is a significant difference between at least one pair of the group means. It does not indicate
what the pair or pairs are significantly different. To find which method is of better performance, a Tukey HSD
test needs to be performed. A Tukey test is interested in examining mean diff erences where any two means that
are greater than HSD are significantly different. From ANOVA results, the mean values of the three input ways:
mouse,1$, and DTW are M1, M2, and M3, respectively.
A post-hoc analysis using a Tukeys HSD testshowed that HSD=13. 18, with M1 - M2 =109.36, M2 -
M3 =108.43, and M3 - M1 =0.93. Thus, the total time of the 1$ recognizer was significantly larger than those
with the other input methods (i.e. no difference was found between the traditional method using the mouse and
the laser pointer using DTW recognizer).
Figure 8: Accuracy of laser pointer with two algorithms.
Second, another one-way ANOVA test is performed in regards to recognition rate (accuracy) of the
laser pointer with the two algorithms: DTW and 1$. The test indicates that there is a significant difference
between the two algorithms (p<0.05). Thus, participants within the DTW method group generated significantly
more accurate gestures than the 1$ recognizer.
Figure 9: Number of errors for all participants in each session.
VI. Conclusion
A real-time system is introduced to track and recognize laser pointer gestures for controlling the
reconstructed 3D image on the large display using preset laser pointer gestures, which can be executed at a
distance from the display. The system includes three fundamental steps: laser spot segmentation, laser spot
tracking and gesture recognition. The conclusion that can be drawn from these tests is that mouse-based
interactions are clearly faster than the laser pointer with 1$ recognizer that shows poor performance with the line
gestures and overall accuracy of %75. The laser pointer with DTW performs about the same as the mouse in the
8. Laser pointer interaction for 3D angiography operation
DOI: 10.9790/0661-17511825 www.iosrjournals.org 25 | Page
completion time with an overall accuracy of %89.6. Considering that, better cameras can improve the total time
that is spent on executing the interactions with the large display, the laser pointer system could be a potential
substitute for the conventional input system in ORs, and the surgeons can perform interactions without needing
help from another staff member.
Future work includes firstly testing the proposed system in a real-world scenario such as the OR.
However, the OR is a very critical place so the accuracy of the laser pointer with DTW should increase.
Secondly, we plan to extend this work to support multiple users with personalized service. Also, we hope to find
out whether other techniques can be useful in ORs, replacing the laser pointing or augmenting it.
References
[1] P. Dias, J. Parracho, J. Cardoso, B. Q. Ferreira, C. Ferreira, and B. S. Santos, “Developing and evaluating two gestural-based virtual
environment navigation methods for large displays,” in Distributed, Ambient, and Pervasive Interactions, pp. 141–151, Springer,
2015.
[2] X. Bi, S.-H. Bae, and R. Balakrishnan, “Walltop: Managing overflowing windows on a large display,” Human{Computer
Interaction, vol. 29, no. 2, pp. 153–203, 2014.
[3] K. Cheng, J. Li, and C. M¨uller-Tomfelde, “Supporting interaction and collaboration on large displays using tablet devices,” in
Proceedings of the International Working Conference onAdvanced Visual Interfaces, pp. 774–775, ACM, 2012.
[4] C. Ardito, P. Buono, M. F. Costabile, and G. Desolda, “Interaction with large displays: A survey,” ACM Computing Surveys
(CSUR), vol. 47, no. 3, p. 46, 2015.
[5] L. Hespanhol, M. Tomitsch, K. Grace, A. Collins, and J. Kay, “Investigating intuitiveness and effectiveness of gestures for free
spatial interaction with large displays,” in Proceedingsof the 2012 International Symposium on Pervasive Displays, p. 6, ACM,
2012.
[6] S. H. Hirano, M. T. Yeganyan, G. Marcu, D. H. Nguyen, L. A. Boyd, and G. R. Hayes, “vsked: evaluation of a system to support
classroom activities for children with autism,” in Proceedings of the SIGCHI Conference on Human Factors in Computing Systems,
pp. 1633–1642, ACM, 2010.
[7] G. Robertson, M. Czerwinski, P. Baudisch, B. Meyers, D. Robbins, G. Smith, and D. Tan, “The large-display user experience,”
Computer Graphics and Applications, IEEE, pp. 44–51, 2005.
[8] K. Cheng and K. Pulo, “Direct interaction with large-scale display systems using infrared laser tracking devices,” in Proceedings of
the Asia-Paci c symposium on Information visualisation-Volume 24, pp. 67–74, Australian Computer Society, Inc., 2003.
[9] D. R. Olsen Jr and T. Nielsen, “Laser pointer interaction,” in Proceedings of the SIGCHIconference on Human factors in computing
systems, pp. 17–22, ACM, 2001.
[10] P. Dey, A. Paul, D. Saha, S. Mukherjee, and A. Nath, “Laser beam operated windows op-eration,” in Communication Systems and
Network Technologies (CSNT), 2012 InternationalConference on, pp. 594–599, IEEE, 2012.
[11] A. Bezerianos and R. Balakrishnan, “The vacuum: facilitating the manipulation of distant objects,” in Proceedings of the SIGCHI
conference on Human factors in computing systems, 361–370, ACM, 2005.
[12] L. Stifelman, A. Elman, and A. Sullivan, “Designing natural speech interactions for the living room,” in CHI'13 Extended Abstracts
on Human Factors in Computing Systems, pp. 1215– 1220, ACM, 2013.
[13] M. Vidal, K. Pfeuffer, A. Bulling, and H. W. Gellersen, “Pursuits: eye-based interaction with moving targets,” in CHI'13 Extended
Abstracts on Human Factors in Computing Systems, 3147–3150, ACM, 2013.
[14] L. Hespanhol, M. Tomitsch, K. Grace, A. Collins, and J. Kay, “Investigating intuitiveness and effectiveness of gestures for free
spatial interaction with large displays,” in Proceedingsof the 2012 International Symposium on Pervasive Displays, p. 6, ACM,
2012.
[15] J. Y. Lee, M. S. Kim, J. S. Kim, and S. M. Lee, “Tangible user interface of digital products in multi-displays,” The International
Journal of Advanced Manufacturing Technology, vol. 59, no. 9-12, pp. 1245–1259, 2012.
[16] R. Jota, M. A. Nacenta, J. A. Jorge, S. Carpendale, and S. Greenberg, “A comparison of ray pointing techniques for very large
displays,” in Proceedings of Graphics Interface 2010, 269–276, Canadian Information Processing Society, 2010.
[17] J.-Y. Oh and W. Stuerzlinger, “Laser pointers as collaborative pointing devices,” in GraphicsInterface, vol. 2002, pp. 141–149,
Citeseer, 2002.
[18] R. Sukthankar, R. G. Stockton, and M. D. Mullin, “Smarter presentations: Exploiting ho-mography in camera-projector systems,” in
Computer Vision, 2001. ICCV 2001. Proceedings.Eighth IEEE International Conference on, vol. 1, pp. 247–253, IEEE, 2001.
[19] B. A. Myers, R. Bhatnagar, J. Nichols, C. H. Peck, D. Kong, R. Miller, and A. C. Long, “Interacting at a distance: measuring the
performance of laser pointers and other devices,” in Proceedings of the SIGCHI conference on Human factors in computing
systems, pp. 33–40, ACM, 2002.
[20] B. Shizuki, T. Hisamatsu, S. Takahashi, and J. Tanaka, “Laser pointer interaction techniques using peripheral areas of screens,” in
Proceedings of the working conference on Advancedvisual interfaces, pp. 95–98, ACM, 2006.
[21] T. Hisamatsu, B. Shizuki, S. Takahashi, and J. Tanaka, “A novel click-free interaction tech-nique for large-screen interfaces,”
APCHI2006, 2006.
[22] X. Bi, Y. Shi, and X. Chen, “Upen: a smart pen-liked device for facilitating interaction on large displays,” in Horizontal Interactive
Human-Computer Systems, 2006. TableTop 2006.First IEEE International Workshop on, pp. 7–pp, IEEE, 2006.
[23] C.-S. Wang, L.-P. Hung, S.-Y. Peng, and L.-C. Cheng, “A laser point interaction system in-tegrating mouse functions,” World
Academy of Science, Engineering and Technology, vol. 67, pp. 657–662, 2010.
[24] J. Shin, S. Kim, and S. Yi, “Development of multi-functional laser pointer mouse through im-age processing,” in Multimedia,
Computer Graphics and Broadcasting, pp. 290–298, Springer, 2011.
[25] S. Sugawara and T. Miki, “A remote lecture system with laser pointer for the internet and broadband networks,” in Applications and
the Internet Workshops, 2004. SAINT 2004 Work-shops. 2004 International Symposium on, pp. 61–66, IEEE, 2004.
[26] F. Ch´avez, F. Fern´andez, R. Alcal´a, J. Alcal´-Fdez, G. Olague, and F. Herrera, “Hybrid laser pointer detection algorithm based on
template matching and fuzzy rule-based systems for domotic control in real home environments,” Applied Intelligence, vol. 36, no.
2, pp. 407– 423, 2012.
[27] O. Barnich and M. Van Droogenbroeck, “Vibe: A universal background subtraction algorithm for video sequences,” Image
Processing, IEEE Transactions on, vol. 20, no. 6, pp. 1709–1724, 2011.