This document presents an automatic algorithm for object recognition and detection based on ASIFT keypoints. The algorithm combines affine scale invariant feature transform (ASIFT) and a region merging algorithm. ASIFT is used to extract keypoints from a training image of the object. These keypoints are then used instead of user markers in a region merging algorithm to recognize and detect the object with full boundary in other images. Experimental results show the method is efficient and accurate at recognizing and detecting objects.
SHORT LISTING LIKELY IMAGES USING PROPOSED MODIFIED-SIFT TOGETHER WITH CONVEN...ijfcstjournal
The paper proposes the modified-SIFT algorithm which will be a modified form of the scale invariant feature transform. The modification consists of considering successive groups of 8 rows of pixel, along the height of the image. These are used to construct 8 bin histograms for magnitude as well as orientation individually. As a result the number of feature descriptors is significantly less (95%) than the standard SIFT approach. Fewer feature descriptor leads to reduced accuracy. This reduction in accuracy is quite drastic when searching for a single (RANK1) image match; however accuracy improves if a band of likely (say tolerance of 10%) images is to be returned. The paper therefore proposes a two-stage-approach where
First Modified-SIFT is used to obtain a shortlisted band of likely images subsequently SIFT is applied within this band to find a perfect match. It may appear that this process is tedious however it provides a significant reduction in search time as compared to applying SIFT on the entire database. The minor reduction in accuracy can be offset by the considerable time gained while searching a large database. The
modified-SIFT algorithm when used in conjunction with a face cropping algorithm can also be used to find a match against disguised images.
3D Reconstruction from Multiple uncalibrated 2D Images of an ObjectAnkur Tyagi
3D reconstruction is the process of capturing the shape and appearance of real objects. In this project we are using passive methods which only use sensors to measure the radiance reflected or emitted by the objects surface to infer its 3D structure.
Object tracking with SURF: ARM-Based platform ImplementationEditor IJCATR
Several algorithms for object tracking, are developed, but our method is slightly different, it’s about how to adapt and implement such algorithms on mobile platform.
We started our work by studying and analyzing feature matching algorithms, to highlight the most appropriate implementation technique for our case.
In this paper, we propose a technique of implementation of the algorithm SURF (Speeded Up Robust Features), for purposes of recognition and object tracking in real time. This is achieved by the realization of an application on a mobile platform such a Raspberry pi, when we can select an image containing the object to be tracked, in the scene captured by the live camera pi. Our algorithm calculates the SURF descriptor for the two images to detect the similarity therebetween, and then matching between similar objects. In the second level, we extend our algorithm to achieve a tracking in real time, all that must respect raspberry pi performances. So, the first thing is setting up all libraries that the raspberry pi need, then adapt the algorithm with card’s performances. This paper presents experimental results on a set of evaluation images as well as images obtained in real time.
EFFECTIVE INTEREST REGION ESTIMATION MODEL TO REPRESENT CORNERS FOR IMAGE sipij
One of the most important steps to describe local features is to estimate the interest region around the feature location to achieve the invariance against different image transformation. The pixels inside the interest region are used to build the descriptor, to represent a feature. Estimating the interest region
around a corner location is a fundamental step to describe the corner feature. But the process is challenging under different image conditions. Most of the corner detectors derive appropriate scales to estimate the region to build descriptors. In our approach, we have proposed a new local maxima-based
interest region detection method. This region estimation method can be used to build descriptors to represent corners. We have performed a comparative analysis to match the feature points using recent corner detectors and the result shows that our method achieves better precision and recall results than
existing methods.
A Detailed Analysis on Feature Extraction Techniques of Panoramic Image Stitc...IJEACS
Image stitching is a technique which is used for attaining a high resolution panoramic image. In this technique, distinct aesthetic images that are imaged from different view and angles are combined together to produce a panoramic image. In the field of computer graphics, photographic and computer vision, Image stitching techniques are considered as current research areas. For obtaining a stitched image it becomes mandatory that one should have the knowledge of geometric relations among multiple image co-ordinate system [1].First, image stitching will be done based on feature key point matches. Final image with seam will be blended with image blending technique. Hence in this paper we are going to address multiple distinct techniques like some invariant features as Scale Invariant Feature Transform and Speeded up Robust Transform and Corner techniques as Harris Corner Detection Technique that are useful in sorting out the issues related with stitching of images.
Improved Characters Feature Extraction and Matching Algorithm Based on SIFTNooria Sukmaningtyas
According to SIFT algorithm does not have the property of affine invariance, and the high
complexity of time and space, it is difficult to apply to real-time image processing for batch image
sequence, so an improved SIFT feature extraction algorithm was proposed in this paper. Firstly, the MSER
algorithm detected the maximally stable extremely regions instead of the DOG operator detected extreme
point, increasing the stability of the characteristics, and reducing the number of the feature descriptor;
Secondly, the circular feature region is divided into eight fan-shaped sub-region instead of 16 square subregion
of the traditional SIFT, and using Gaussian function weighted gradient information field to construct
the new SIFT features descriptor. Compared with traditional SIFT algorithm, The experimental results
showed that the algorithm not only has translational invariance, scale invariance and rotational invariance,
but also has affine invariance and faster speed that meet the requirements of real-time image processing
applications.
SHORT LISTING LIKELY IMAGES USING PROPOSED MODIFIED-SIFT TOGETHER WITH CONVEN...ijfcstjournal
The paper proposes the modified-SIFT algorithm which will be a modified form of the scale invariant feature transform. The modification consists of considering successive groups of 8 rows of pixel, along the height of the image. These are used to construct 8 bin histograms for magnitude as well as orientation individually. As a result the number of feature descriptors is significantly less (95%) than the standard SIFT approach. Fewer feature descriptor leads to reduced accuracy. This reduction in accuracy is quite drastic when searching for a single (RANK1) image match; however accuracy improves if a band of likely (say tolerance of 10%) images is to be returned. The paper therefore proposes a two-stage-approach where
First Modified-SIFT is used to obtain a shortlisted band of likely images subsequently SIFT is applied within this band to find a perfect match. It may appear that this process is tedious however it provides a significant reduction in search time as compared to applying SIFT on the entire database. The minor reduction in accuracy can be offset by the considerable time gained while searching a large database. The
modified-SIFT algorithm when used in conjunction with a face cropping algorithm can also be used to find a match against disguised images.
3D Reconstruction from Multiple uncalibrated 2D Images of an ObjectAnkur Tyagi
3D reconstruction is the process of capturing the shape and appearance of real objects. In this project we are using passive methods which only use sensors to measure the radiance reflected or emitted by the objects surface to infer its 3D structure.
Object tracking with SURF: ARM-Based platform ImplementationEditor IJCATR
Several algorithms for object tracking, are developed, but our method is slightly different, it’s about how to adapt and implement such algorithms on mobile platform.
We started our work by studying and analyzing feature matching algorithms, to highlight the most appropriate implementation technique for our case.
In this paper, we propose a technique of implementation of the algorithm SURF (Speeded Up Robust Features), for purposes of recognition and object tracking in real time. This is achieved by the realization of an application on a mobile platform such a Raspberry pi, when we can select an image containing the object to be tracked, in the scene captured by the live camera pi. Our algorithm calculates the SURF descriptor for the two images to detect the similarity therebetween, and then matching between similar objects. In the second level, we extend our algorithm to achieve a tracking in real time, all that must respect raspberry pi performances. So, the first thing is setting up all libraries that the raspberry pi need, then adapt the algorithm with card’s performances. This paper presents experimental results on a set of evaluation images as well as images obtained in real time.
EFFECTIVE INTEREST REGION ESTIMATION MODEL TO REPRESENT CORNERS FOR IMAGE sipij
One of the most important steps to describe local features is to estimate the interest region around the feature location to achieve the invariance against different image transformation. The pixels inside the interest region are used to build the descriptor, to represent a feature. Estimating the interest region
around a corner location is a fundamental step to describe the corner feature. But the process is challenging under different image conditions. Most of the corner detectors derive appropriate scales to estimate the region to build descriptors. In our approach, we have proposed a new local maxima-based
interest region detection method. This region estimation method can be used to build descriptors to represent corners. We have performed a comparative analysis to match the feature points using recent corner detectors and the result shows that our method achieves better precision and recall results than
existing methods.
A Detailed Analysis on Feature Extraction Techniques of Panoramic Image Stitc...IJEACS
Image stitching is a technique which is used for attaining a high resolution panoramic image. In this technique, distinct aesthetic images that are imaged from different view and angles are combined together to produce a panoramic image. In the field of computer graphics, photographic and computer vision, Image stitching techniques are considered as current research areas. For obtaining a stitched image it becomes mandatory that one should have the knowledge of geometric relations among multiple image co-ordinate system [1].First, image stitching will be done based on feature key point matches. Final image with seam will be blended with image blending technique. Hence in this paper we are going to address multiple distinct techniques like some invariant features as Scale Invariant Feature Transform and Speeded up Robust Transform and Corner techniques as Harris Corner Detection Technique that are useful in sorting out the issues related with stitching of images.
Improved Characters Feature Extraction and Matching Algorithm Based on SIFTNooria Sukmaningtyas
According to SIFT algorithm does not have the property of affine invariance, and the high
complexity of time and space, it is difficult to apply to real-time image processing for batch image
sequence, so an improved SIFT feature extraction algorithm was proposed in this paper. Firstly, the MSER
algorithm detected the maximally stable extremely regions instead of the DOG operator detected extreme
point, increasing the stability of the characteristics, and reducing the number of the feature descriptor;
Secondly, the circular feature region is divided into eight fan-shaped sub-region instead of 16 square subregion
of the traditional SIFT, and using Gaussian function weighted gradient information field to construct
the new SIFT features descriptor. Compared with traditional SIFT algorithm, The experimental results
showed that the algorithm not only has translational invariance, scale invariance and rotational invariance,
but also has affine invariance and faster speed that meet the requirements of real-time image processing
applications.
Real-time Moving Object Detection using SURFiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
This project is to retrieve the similar geographic images from the dataset based on the features extracted.
Retrieval is the process of collecting the relevant images from the dataset which contains more number of
images. Initially the preprocessing step is performed in order to remove noise occurred in input image with
the help of Gaussian filter. As the second step, Gray Level Co-occurrence Matrix (GLCM), Scale Invariant
Feature Transform (SIFT), and Moment Invariant Feature algorithms are implemented to extract the
features from the images. After this process, the relevant geographic images are retrieved from the dataset
by using Euclidean distance. In this, the dataset consists of totally 40 images. From that the images which
are all related to the input image are retrieved by using Euclidean distance. The approach of SIFT is to
perform reliable recognition, it is important that the feature extracted from the training image be
detectable even under changes in image scale, noise and illumination. The GLCM calculates how often a
pixel with gray level value occurs. While the focus is on image retrieval, our project is effectively used in
the applications such as detection and classification.
Two Dimensional Image Reconstruction Algorithmsmastersrihari
Convolution Back-Projection (CBP) Algorithm was used for the reconstruction of the image. The performance was compared by implementing the algorithm by using RAM- LAK filter, Shepp- Logan filter and also No filter being used.
Edge detection is still difficult task in the image processing field. In this paper we implemented fuzzy techniques for detecting edges in the image. This algorithm also works for medical images. In this paper we also explained about Fuzzy inference system, which is more robust to contrast and lighting variations.
Matching algorithm performance analysis for autocalibration method of stereo ...TELKOMNIKA JOURNAL
Stereo vision is one of the interesting research topics in the computer vision field. Two cameras are used to generate a disparity map, resulting in the depth estimation. Camera calibration is the most important step in stereo vision. The calibration step is used to generate an intrinsic parameter of each camera to get a better disparity map. In general, the calibration process is done manually by using a chessboard pattern, but this process is an exhausting task. Self-calibration is an important ability required to overcome this problem. Self-calibration required a robust and good matching algorithm to find the key feature between images as reference. The purpose of this paper is to analyze the performance of three matching algorithms for the autocalibration process. The matching algorithms used in this research are SIFT, SURF, and ORB. The result shows that SIFT performs better than other methods.
Leader Follower Formation Control of Ground Vehicles Using Dynamic Pixel Coun...ijma
This paper deals with leader-follower formations of non-holonomic mobile robots, introducing a formation
control strategy based on pixel counts using a commercial grade electro optics camera. Localization of the
leader for motions along line of sight as well as the obliquely inclined directions are considered based on
pixel variation of the images by referencing to two arbitrarily designated positions in the image frames.
Based on an established relationship between the displacement of the camera movement along the viewing
direction and the difference in pixel counts between reference points in the images, the range and the angle
estimate between the follower camera and the leader is calculated. The Inverse Perspective Transform is
used to account for non linear relationship between the height of vehicle in a forward facing image and its
distance from the camera. The formulation is validated with experiments.
Comparative Analysis of Common Edge Detection Algorithms using Pre-processing...IJECEIAES
Edge detection is the process of segmenting an image by detecting discontinuities in brightness. Several standard segmentation methods have been widely used for edge detection. However, due to inherent quality of images, these methods prove ineffective if they are applied without any preprocessing. In this paper, an image pre-processing approach has been adopted in order to get certain parameters that are useful to perform better edge detection with the standard edge detection methods. The proposed preprocessing approach involves median filtering to reduce the noise in image and then edge detection technique is carried out. Finally, Standard edge detection methods can be applied to the resultant pre-processing image and its Simulation results are show that our pre-processed approach when used with a standard edge detection method enhances its performance.
Feature Matching using SIFT algorithm; co-authored presentation on Photogrammetry studio by Sajid Pareeth, Gabriel Vincent Sanya, Sonam Tashi and Michael Mutale
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...CSCJournals
This framework is a hand movement capturing method which could be done in three different depth levels. The algorithm has the capability of capturing and identifying when the hand is moving up, down, right and left. From these captured movements four signals could be generated. Moreover, when these hand movements are done, 15cm-75cm, 75cm-100cm, 100cm- 200cm from the camera (3 depth levels), twelve different signals could be generated. These generated signals could be used for applications such as game controlling (gaming).The existing method uses an object area based method for depth analysis. The results of the proposed work shows it has high accuracy compared to the existing method when tested for depth analysis.
Based on correlation coefficient in image matchingIJRES Journal
With the development of image technology, the application of image technology in industrial
manufacturing become more and more widely. Image matching is animportantbranch of the image processing,
and also it is very important in the process of industrial detection. The main purpose of this paper is aiming at
the traditional image correlation matching algorithm in the application of detection in industrial, which based
on similarity matching principle. This paper is going to discuss the similarity of the image matching algorithm
on the application of detection in the connector. This paper improves the algorithm to speed up the detection
velocity. In some specific circumstance, while correlation coefficient method is more simple and easy to use, the
software development cycle will be short.
ANALYSIS OF INTEREST POINTS OF CURVELET COEFFICIENTS CONTRIBUTIONS OF MICROS...sipij
This paper focuses on improved edge model based on Curvelet coefficients analysis. Curvelet transform is
a powerful tool for multiresolution representation of object with anisotropic edge. Curvelet coefficients
contributions have been analyzed using Scale Invariant Feature Transform (SIFT), commonly used to study
local structure in images. The permutation of Curvelet coefficients from original image and edges image
obtained from gradient operator is used to improve original edges. Experimental results show that this
method brings out details on edges when the decomposition scale increases.
Geometric wavelet transform for optical flow estimation algorithmijcga
This paper described an algorithm for computing the optical flow (OF) vector of a moving objet in a video sequence based on geometric wavelet transform (GWT). This method tries to calculate the motion between two successive frames by using a GWT. It consists to project the OF vectors on a basis of geometric wavelet. Using GWT for OF estimation has been attracting much attention. This approach takes advantage of the geometric wavelet filter property and requires only two frames. This algorithm is fast and able to estimate the OF with a low-complexity. The technique is suitable for video compression, and can be used for stereo vision and image registration.
An Assessment of Image Matching Algorithms in Depth EstimationCSCJournals
Computer vision is often used with mobile robot for feature tracking, landmark sensing, and obstacle detection. Almost all high-end robotics systems are now equipped with pairs of cameras arranged to provide depth perception. In stereo vision application, the disparity between the stereo images allows depth estimation within a scene. Detecting conjugate pair in stereo images is a challenging problem known as the correspondence problem. The goal of this research is to assess the performance of SIFT, MSER, and SURF, the well known matching algorithms, in solving the correspondence problem and then in estimating the depth within the scene. The results of each algorithm are evaluated and presented. The conclusion and recommendations for future works, lead towards the improvement of these powerful algorithms to achieve a higher level of efficiency within the scope of their performance.
Real-time Moving Object Detection using SURFiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
This project is to retrieve the similar geographic images from the dataset based on the features extracted.
Retrieval is the process of collecting the relevant images from the dataset which contains more number of
images. Initially the preprocessing step is performed in order to remove noise occurred in input image with
the help of Gaussian filter. As the second step, Gray Level Co-occurrence Matrix (GLCM), Scale Invariant
Feature Transform (SIFT), and Moment Invariant Feature algorithms are implemented to extract the
features from the images. After this process, the relevant geographic images are retrieved from the dataset
by using Euclidean distance. In this, the dataset consists of totally 40 images. From that the images which
are all related to the input image are retrieved by using Euclidean distance. The approach of SIFT is to
perform reliable recognition, it is important that the feature extracted from the training image be
detectable even under changes in image scale, noise and illumination. The GLCM calculates how often a
pixel with gray level value occurs. While the focus is on image retrieval, our project is effectively used in
the applications such as detection and classification.
Two Dimensional Image Reconstruction Algorithmsmastersrihari
Convolution Back-Projection (CBP) Algorithm was used for the reconstruction of the image. The performance was compared by implementing the algorithm by using RAM- LAK filter, Shepp- Logan filter and also No filter being used.
Edge detection is still difficult task in the image processing field. In this paper we implemented fuzzy techniques for detecting edges in the image. This algorithm also works for medical images. In this paper we also explained about Fuzzy inference system, which is more robust to contrast and lighting variations.
Matching algorithm performance analysis for autocalibration method of stereo ...TELKOMNIKA JOURNAL
Stereo vision is one of the interesting research topics in the computer vision field. Two cameras are used to generate a disparity map, resulting in the depth estimation. Camera calibration is the most important step in stereo vision. The calibration step is used to generate an intrinsic parameter of each camera to get a better disparity map. In general, the calibration process is done manually by using a chessboard pattern, but this process is an exhausting task. Self-calibration is an important ability required to overcome this problem. Self-calibration required a robust and good matching algorithm to find the key feature between images as reference. The purpose of this paper is to analyze the performance of three matching algorithms for the autocalibration process. The matching algorithms used in this research are SIFT, SURF, and ORB. The result shows that SIFT performs better than other methods.
Leader Follower Formation Control of Ground Vehicles Using Dynamic Pixel Coun...ijma
This paper deals with leader-follower formations of non-holonomic mobile robots, introducing a formation
control strategy based on pixel counts using a commercial grade electro optics camera. Localization of the
leader for motions along line of sight as well as the obliquely inclined directions are considered based on
pixel variation of the images by referencing to two arbitrarily designated positions in the image frames.
Based on an established relationship between the displacement of the camera movement along the viewing
direction and the difference in pixel counts between reference points in the images, the range and the angle
estimate between the follower camera and the leader is calculated. The Inverse Perspective Transform is
used to account for non linear relationship between the height of vehicle in a forward facing image and its
distance from the camera. The formulation is validated with experiments.
Comparative Analysis of Common Edge Detection Algorithms using Pre-processing...IJECEIAES
Edge detection is the process of segmenting an image by detecting discontinuities in brightness. Several standard segmentation methods have been widely used for edge detection. However, due to inherent quality of images, these methods prove ineffective if they are applied without any preprocessing. In this paper, an image pre-processing approach has been adopted in order to get certain parameters that are useful to perform better edge detection with the standard edge detection methods. The proposed preprocessing approach involves median filtering to reduce the noise in image and then edge detection technique is carried out. Finally, Standard edge detection methods can be applied to the resultant pre-processing image and its Simulation results are show that our pre-processed approach when used with a standard edge detection method enhances its performance.
Feature Matching using SIFT algorithm; co-authored presentation on Photogrammetry studio by Sajid Pareeth, Gabriel Vincent Sanya, Sonam Tashi and Michael Mutale
An Enhanced Computer Vision Based Hand Movement Capturing System with Stereo ...CSCJournals
This framework is a hand movement capturing method which could be done in three different depth levels. The algorithm has the capability of capturing and identifying when the hand is moving up, down, right and left. From these captured movements four signals could be generated. Moreover, when these hand movements are done, 15cm-75cm, 75cm-100cm, 100cm- 200cm from the camera (3 depth levels), twelve different signals could be generated. These generated signals could be used for applications such as game controlling (gaming).The existing method uses an object area based method for depth analysis. The results of the proposed work shows it has high accuracy compared to the existing method when tested for depth analysis.
Based on correlation coefficient in image matchingIJRES Journal
With the development of image technology, the application of image technology in industrial
manufacturing become more and more widely. Image matching is animportantbranch of the image processing,
and also it is very important in the process of industrial detection. The main purpose of this paper is aiming at
the traditional image correlation matching algorithm in the application of detection in industrial, which based
on similarity matching principle. This paper is going to discuss the similarity of the image matching algorithm
on the application of detection in the connector. This paper improves the algorithm to speed up the detection
velocity. In some specific circumstance, while correlation coefficient method is more simple and easy to use, the
software development cycle will be short.
ANALYSIS OF INTEREST POINTS OF CURVELET COEFFICIENTS CONTRIBUTIONS OF MICROS...sipij
This paper focuses on improved edge model based on Curvelet coefficients analysis. Curvelet transform is
a powerful tool for multiresolution representation of object with anisotropic edge. Curvelet coefficients
contributions have been analyzed using Scale Invariant Feature Transform (SIFT), commonly used to study
local structure in images. The permutation of Curvelet coefficients from original image and edges image
obtained from gradient operator is used to improve original edges. Experimental results show that this
method brings out details on edges when the decomposition scale increases.
Geometric wavelet transform for optical flow estimation algorithmijcga
This paper described an algorithm for computing the optical flow (OF) vector of a moving objet in a video sequence based on geometric wavelet transform (GWT). This method tries to calculate the motion between two successive frames by using a GWT. It consists to project the OF vectors on a basis of geometric wavelet. Using GWT for OF estimation has been attracting much attention. This approach takes advantage of the geometric wavelet filter property and requires only two frames. This algorithm is fast and able to estimate the OF with a low-complexity. The technique is suitable for video compression, and can be used for stereo vision and image registration.
An Assessment of Image Matching Algorithms in Depth EstimationCSCJournals
Computer vision is often used with mobile robot for feature tracking, landmark sensing, and obstacle detection. Almost all high-end robotics systems are now equipped with pairs of cameras arranged to provide depth perception. In stereo vision application, the disparity between the stereo images allows depth estimation within a scene. Detecting conjugate pair in stereo images is a challenging problem known as the correspondence problem. The goal of this research is to assess the performance of SIFT, MSER, and SURF, the well known matching algorithms, in solving the correspondence problem and then in estimating the depth within the scene. The results of each algorithm are evaluated and presented. The conclusion and recommendations for future works, lead towards the improvement of these powerful algorithms to achieve a higher level of efficiency within the scope of their performance.
MULTIPLE REGION OF INTEREST TRACKING OF NON-RIGID OBJECTS USING DEMON'S ALGOR...cscpconf
In this paper we propose an algorithm for tracking multiple ROI (region of interest) undergoing non-rigid transformations. Demon's algorithm based on the idea of Maxwell's demon, has been applied here to estimate the displacement field for tracking of multiple ROI. This algorithm works on pixel intensities of the sequence of images thus making it suitable for tracking objects/regions undergoing non-rigid transformations. We have incorporated a pyramid-based approach for demon's algorithm computations of displacement field, which leads to significant reduction in the convergence speed and improvement in the accuracy. This algorithm is applied for tracking non-rigid objects in laproscopy videos which would aid surgeons in Minimal Invasive Surgery (MIS).
Multiple region of interest tracking of non rigid objects using demon's algor...csandit
In this paper we propose an algorithm for tracking multiple ROI (region of interest) undergoing
non-rigid transformations. Demon's algorithm based on the idea of Maxwell's demon, has been
applied here to estimate the displacement field for tracking of multiple ROI. This algorithm
works on pixel intensities of the sequence of images thus making it suitable for tracking
objects/regions undergoing non-rigid transformations. We have incorporated a pyramid-based
approach for demon's algorithm computations of displacement field, which leads to significant
reduction in the convergence speed and improvement in the accuracy. This algorithm is applied
for tracking non-rigid objects in laproscopy videos which would aid surgeons in Minimal
Invasive Surgery (MIS).
Automatic rectification of perspective distortion from a single image using p...ijcsa
Perspective distortion occurs due to the perspective projection of 3D scene on a 2D surface. Correcting the distortion of a single image without losing any desired information is one of the challenging task in the field of Computer Vision. We consider the problem of estimating perspective distortion from a single still image of an unstructured environment and to make perspective correction which is both quantitatively accurate as well as visually pleasing. Corners are detected based on the orientation of the image. A method based on plane homography and transformation is used to make perspective correction. The algorithm infers frontier information directly from the images, without any reference objects or prior knowledge of the camera parameters. The frontiers are detected using geometric context based segmentation. The goal of this paper is to present a framework providing fully automatic and fast perspective correction.
Performance Evaluation of Image Edge Detection Techniques CSCJournals
The success of an image recognition procedure is related to the quality of the edges marked. The
aim of this research is to investigate and evaluate edge detection techniques when applied to
noisy images at different scales. Sobel, Prewitt, and Canny edge detection algorithms are
evaluated using artificially generated images and comparison criteria: edge quality (EQ) and map
quality (MQ). The results demonstrated that the use of these criteria can be utilized as an aid for
further analysis and arbitration to find the best edge detector for a given image.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Dense Visual Odometry Using Genetic AlgorithmSlimane Djema
Our work aims to estimate the camera motion mounted on the head of a mobile robot or a moving object from RGB-D images in a static scene. The problem of motion estimation is transformed into a nonlinear least squares function. Methods for solving such problems are iterative. Various classic methods gave an iterative solution by linearizing this function. We can also use the metaheuristic optimization method to solve this problem and improve results. In this paper, a new algorithm is developed for visual odometry using a sequence of RGB-D images. This algorithm is based on a genetic algorithm. The proposed iterative genetic algorithm searches using particles to estimate the optimal motion and then compares it to the traditional methods. To evaluate our method, we use the root mean square error to compare it with the based energy method and another metaheuristic method. We prove the efficiency of our innovative algorithm on a large set of images.
Gait Based Person Recognition Using Partial Least Squares Selection Scheme ijcisjournal
The variations of viewing angle and intra-class of human beings have great impact on gait recognition
systems. This work represents an Arbitrary View Transformation Model (AVTM) for recognizing the gait.
Gait energy image (GEI) based gait authentication is effective approach to address the above problem, the
method establishes an AVTM based on principle component analysis (PCA). Feature selection (FS) is
performed using Partial least squares (PLS) method. The comparison of the AVTM PLS method with the
existing methods shows significant advantages in terms of observing angle variation, carrying and attire
changes. Experiments evaluated over CASIA gait database, shows that the proposed method improves the
accuracy of recognition compared to the other existing methods.
Touchless Palmprint Verification using Shock Filter, SIFT, I-RANSAC, and LPD iosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Using Generic Image Processing Operations to Detect a Calibration GridJan Wedekind
Camera calibration is an important problem in 3D computer vision. The problem of determining the camera parameters has been studied extensively. However the algorithms for determining the required correspondences are either semi-automatic (i.e. they require user interaction) or they involve difficult to implement custom algorithms.
We present a robust algorithm for detecting the corners of a calibration grid and assigning the correct correspondences for calibration . The solution is based on generic image processing operations so that it can be implemented quickly. The algorithm is limited to distortion-free cameras but it could potentially be extended to deal with camera distortion as well. We also present a corner detector based on steerable filters. The corner detector is particularly suited for the problem of detecting the corners of a calibration grid.
- See more at: http://figshare.com/articles/Using_Generic_Image_Processing_Operations_to_Detect_a_Calibration_Grid/696880#sthash.EG8dWyTH.dpuf
Efficient 3D stereo vision stabilization for multi-camera viewpointsjournalBEEI
In this paper, an algorithm is developed in 3D Stereo vision to improve image stabilization process for multi-camera viewpoints. Finding accurate unique matching key-points using Harris Laplace corner detection method for different photometric changes and geometric transformation in images. Then improved the connectivity of correct matching pairs by minimizing
the global error using spanning tree algorithm. Tree algorithm helps to stabilize randomly positioned camera viewpoints in linear order. The unique matching key-points will be calculated only once with our method.
Then calculated planar transformation will be applied for real time video rendering. The proposed algorithm can process more than 200 camera viewpoints within two seconds.
Performance of Efficient Closed-Form Solution to Comprehensive Frontier Exposureiosrjce
IOSR Journal of Electronics and Communication Engineering(IOSR-JECE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of electronics and communication engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in electronics and communication engineering. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Similar to An automatic algorithm for object recognition and detection based on asift keypoints kunal (20)
An automatic algorithm for object recognition and detection based on asift keypoints kunal
1. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
DOI : 10.5121/sipij.2012.3503 29
AN AUTOMATIC ALGORITHM FOR OBJECT
RECOGNITION AND DETECTION BASED ON ASIFT
KEYPOINTS
Reza Oji
Department of Computer Engineering and IT, Shiraz University
Shiraz, Iran
oji.reza@gmail.com
ABSTRACT
Object recognition is an important task in image processing and computer vision. This paper presents a
perfect method for object recognition with full boundary detection by combining affine scale invariant
feature transform (ASIFT) and a region merging algorithm. ASIFT is a fully affine invariant algorithm that
means features are invariant to six affine parameters namely translation (2 parameters), zoom, rotation
and two camera axis orientations. The features are very reliable and give us strong keypoints that can be
used for matching between different images of an object. We trained an object in several images with
different aspects for finding best keypoints of it. Then, a robust region merging algorithm is used to
recognize and detect the object with full boundary in the other images based on ASIFT keypoints and a
similarity measure for merging regions in the image. Experimental results show that the presented method
is very efficient and powerful to recognize the object and detect it with high accuracy.
KEYWORDS
Object recognition, keypoint, affine invariant, region merging algorithm, ASIFT.
1. INTRODUCTION
Extracting the points from an image that can give best define from an object in image namely
keypoints is very important and valuable. These points have many applications in image
processing like object detection, object and shape recognition, image registation and object
tracking. By extracting the keypoints, we can use them for finding objects in the other images.
Detect and recognize the object by using the keypoints and a segmentation algorithm is very
accurate because if the keypoints are correctly identified, they achieve the best information from
the image. ASIFT is a fully affine invariant method with respect to six parameters of affine
transform [1,2], wherease the previous method SIFT was invariant with respect to four parameters
namely translation, rotation and change in scale (zoom) [3]. ASIFT cover two other parameters
namely longitude and latitude angle that are relevant to camera axis orientation. It means that
ASIFT is more effective for our goal and can be more robust in the changes of images.
Many image segmentation and object recognition algorithms have been presented, each having its
own specifications [4,5]. Some of these algorithm are interactive image segmentation based on
region merging [6,7]. One of them is a powerful algorithm [8] for detecting object and its
boundary but it is not an automatic algorithm and has a problem. In this algorithm the users must
indicate some of locations and regions of the background and object to run the algorithm.
2. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
30
In this paper, we combined ASIFT results and a region merging algorithm to recognize objects in
images and detect them with full boundary. The presented algorithm does not have the stated
problem in region merging algorithm. It means that, the algorithm does not need to indicate
regions by user. We use the best keypoints of object that has been obtained from ASIFT results
and apply them into the image. Therefore, the method will be an automatic algorithm and will not
need marks by users and the achieved keypoints from ASIFT have been replaced with them.
Lowe presented an object recognition algorithm [9] based on SIFT, but object recognition and
detection algorithm with full boundary by using ASITF, has not yet been presented. Therefore, at
this time we have an automatic algorithm for object recognition and full detection by using the
obtained keypoints from ASIFT algorithm. These keypoints are very efficient and give us the best
information of object to find it in the other images.
2. THE ASIFT ALGORITHM
ASIFT is an improved algorithm from SIFT. SIFT has been presented by Lowe (1994) where
SIFT is invariant to four of the six parameters of affine transform. ASIFT simulates all image
views obtained by the longitude and the latatude angle (varying the two camera axis orientation
parameters) and then applies the other parameters by using the SIFT algorithm. There is a new
notion called transition tilt that is designed to quantify the amount of tilt between two such
images. In continuation, major steps of the algorithm are described.
2.1. The Affine Camera Model
Image distortions coming from viewpoint changes can be modeled by affine planar transforms.
Any affine map A is defined:
−
−
=Α
ϕϕ
ϕϕ
ψψ
ψψ
λ cossin
sincos
10
0
cossin
sincos t
(1)
Which we note A = λ R1(ψ)Tt R2(φ), where λ > 0, λt is the deteminant of A, Ri are rotations, φ ϵ
[0,180), and Tt is a tilt namely a diagonal matrix with first eigenvalue t > 1 and the second one
equal to 1. Figure1, shows a camera motion interpretation of (1). θ and φ = arccos 1/ t are
respectively the camera optical axis longitude and latitude, λ corresponds to the zoom and a third
angle ψ parameterizes the camera spin.
Figure 1. Camera motion interpretation
According to up preamble, each image should be transformed by simulating all possible affine
distortions caused by the changes of camera optical axis orientations from a frontal position.
These distortions depend upon two parameters, the longitude and the latitude. rotations and tilts
are performed for a finite number of angles to decrease complexity and time computation. The
sampling steps of these parameters ensuring that the simulated images keep close to any other
possible view generated by other values of φ and θ.
3. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
31
After this process, the SIFT algorithm will be applied in all simulated images. The other steps are
relevant to SIFT.
2.2. Scale Space Extremum Detection
The next step is extracting keypoints that are invariant to changes of scale. We need to search for
stable features in all possible changes. In previous research it has been shown that for this task,
the Gaussian function is the only possible scale-space kernel. A function, L(x,y,σ), is the scale-
space of an image that is obtained from the convolution of an input image, I(x,y), with a variable
scale-space Gaussian function, G(x,y,σ). For efficiently detect stable keypoint locations, Lowe
proposed using the scale-space extremum in difference-of-Gaussian (DOG) function, D(x,y,σ).
Extremum computed by the difference of two nearby scales separated by a constant factor K (2):
, (2)
This process repeats in several octaves. In each octave, the initial image is repeatedly convolved
with Gaussians to produce the set of scale space images. Adjacent Gaussian images are subtracted
to produce the difference-of-Gaussian images. After each octave, the Gaussian images are down-
sampled by a factor of 2, and the process repeated. Figure 2, shows a visual representation from
this process.
Figure 2. Visual representation of DOG in octaves
Each sample point (pixel) is compared with its neighbors according to their intensities for finding
out whether is smaller or larger than neighbors. For more accuracy, each pixel will be checked
with the eight closest neighbors in image location and nine neighbors in the scale above and
below (Figure 3). If the point is an extremum against all 26 neighbors, is selected as candidate
keypoint. The cost of this comparison is reasonably low because most sample point will be
eliminated at first few checks.
( ) ( )2 2 2
2
2
1
, ,
2
x y
G x y e
σ
σ
πσ
− +
=
( ) ( ) ( )( ) ( )
( ) ( )
, , , , , , ,
, , , ,
D x y G x y k G x y I x y
L x y k L x y
σ σ σ
σ σ
= − ∗
= −
4. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
32
Figure 3. Local extremum detection of DOG
2.3. Accurate Keypoint Localization
For each candidate keypoint interpolation of nearby data is used to accurately determine its
position. Then, many keypoints that are unstable and sensitive to noise, like the points with low
contrast and the points on the edge, will be eliminated. (Figure 4)
Figure 4. (a) Extremum of DOG across scales. (b) Remaining keypoints after removal of low
contrast points. (c) Remaining keypoints after removal of edge responses.
2.4. Assigning an Orientation
The next step is assigning an orientation for each keypoint. By this step, the keypoint descriptor
[10] can be represented relative to this orientation and therefore get invariance to image rotation.
For each Gaussian smoothed image sample, points in regions around keypoint are selected and
magnitude, m, and orientations, θ, of gradiant are calculated (3):
(3)
Then, created weighted (magnitude + gaussian) histogram of local gradiant directions computed
at selected scale. Histogram is formed by quantizing the orientations into 36 bins to covering 360
degree range of orinetations. The highest peak in the histogram is detected where peaks in the
orientation histogram correspond to dominant directions of local gradiant. (Figure 5)
22
))1,()1,(()),1(),1((),( −−++−−+= yxLyxLyxLyxLyxm
))),1(),1(/())1,()1,(((tan),( 1
yxLyxLyxLyxLyx −−+−−+= −
θ
5. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
33
Figure 5. Visual representation of orientation histogram.
(the histogram is shown in 7 bins)
2.5. Keypoint Descriptor
The previous operations have assigned location, scale, and orientation to each keypoint and
provided invariance to these parameters. Remaining goals are to define a keypoint descriptor [11]
for the local image regions and reduce the effects of illumination changes.
The descriptor is based on 16×16 samples that the keypoint is in the center of. Samples are
divided into 4×4 subregions in 8 directions around keypoint. Magnitude of each point is weighted
and gives less weight to gradients far from keypoint (Figure 6). Therefore, feature vector
dimensional is 128 (4×4×8).
Figure 6. A keypoint descriptor
Finally, vector normalization is applied. The vector is normalized to unit length. A change in
image contrast in which each pixel value is multiplied by a constant will multiply gradients by the
same constant. Contrast change will be canceled by vector normalization and brightness change
in which a constant is added to each image pixel will not affect the gradient values, as they are
computed from pixel differences.
Now we can find keypoints from an image in the other images and match them together. One
image is the training sample of what we are looking for and the other image is the world picture
that might contain instances of the training sample. Both images have features associated with
them across different octaves. Keypoints in all octaves in one image independently match with all
keypoints in all octaves in other image. Features will be matched by using nearest neighbor
algorithm. The nearest neighbor is defined as the keypoint with minimum Eculidean distance for
the invariant descriptor vector as described upside. Also to solve the problem of features that have
no correct match due to some reason like background noise or clutter, a threshold at 0.8 is chosen
for ratio of closest nearest neighbor with second closest nearest neighbor that obtained
6. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
34
experimentally . If the distance is more than 0.8, then the algorithm does not match keypoints
together.
3. PROPOSED OBJECT RECOGNITION AND DETECTION METHOD
For each object, we train it by ASIFT algorithm in several images with different scales and
viewpoints to find best keypoints to recognize desire object in the other images. Then, by using
these keypoints and a region merging algorithm the object will be detected. A maximal similarity
region mering segmentation algorithm was propoesd by Ning. This algorithm is based on a region
merging method with using initial segmentation of mean shift [12] and user markers. User marks
part of object and background to help the region merging process by maximal similarity. The
non-marker background regions will be automatically merged while the non-marker object
regions will be identified and avoided from being merged with background. It can extract the
object from complex scenes but there is a weakness that the algorithm needs user marks, hence,
the algorithm is not automatic. In our algorithm, we use this region merging but without user
marks. We propose an efficient method for object recognition and detection by combining ASIFT
keypoints (instead of user marks) with an automatic segmentation based on region merging which
can detect object with full boundary.
3.1. Region Merging Based on Maximal Similarity
An initial segmentation is required to partition image into regions for region merging in further
steps. We use the mean shift segmentation software namely the EDISON System [13] for initial
segmentation because it can preserve the boundries well and has high speed. Any other low level
segmentation like super-pixel and watershed can be used too. Please refer to [14,15] for more
information about mean shift segmentation.
There is many small region after initial segmentation. These regions represent using color
histogram that is an effective descriptor because different regions from the same object often
have high similarity in color whereas there have variation in other aspects like size and shape.
The color histogram computed by RGB color space. We uniformly quantize each color chanel
into 8 levels. Therefore, the feature space is 8×8×8=512 and the histogram of each region is
computed in 512 bins. The normalized histogram of a region X denote by HistX. Now we want to
merge the regions by their color histogram that can extract the desired object.
When we applied ASIFT keypoints in the image, some regions in image that are relevant to object
in image are compoesd of keypoints and the others are not. Here, an important issue is how to
define the similarity between the regions inclusive keypoints with the regions without any point
so that the similar regions can be merged. We use the Euclidean distance for this target because
that is a well known goodness of fit statistical metric and simple.Therefore, ∆(X,Y) is a similarity
measure between two regions X and Y based on the Euclidean distance (4):
( ) ( )∑=
−=∆
512
1
2
,
e
e
Y
e
X HistHistYX (4)
Where HistX and HistY are the normalized histograms of X and Y. Also, The superscript m is mth
member of the histograms. Lower Euclidean distance between X and Y means that the similarity
between them is higher.
7. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
35
3.2. The Merging Process
In the merging process we have three regions in the image with diferent lables: the object regions
denote by RO that are identified by obtained keypoints from training ASIFT algorithm, the regions
around the images denote by RB that usually are not inclusive objects, therefore, we cover all
around the images as initial background regions to help and start the merging process and the
third regions, the regions without any sign denote by N.
The merging rule is defined in continue (5). Let Y be an adjacent region of X and SY is the set of
Y’s adjacent region that X is one of them. The similarity between Y and SY, i.e. ∆(Y,SY), is
calculated. We will merge X and Y together if and only if the similarity between them is the
maximum among all the similarities ∆(Y,SY).
Merge X and Y if ∆(X,Y) = max ∆(Y,SY) (5)
The main strategy is to keep object regions from merging and merge background regions as many
as possible. In the first stage, we try to merge background regions with their adjacent regions.
Each region B ϵ RB will be merged into adjacent region if the merging rule is be satisfied. If the
merging accured, the new region has the same label as region B. This stage will be repeated
iteratively and in each iteration RB and N will be updated. Obviously, RB expands and N shrink
due to merging process. The process in this stage stops when background regions RB can not find
any region for new merging.
After this stage, still, some background regions are, that can not be merged due to merging rule.
In next stage we will focus on the remaining regions in N from the first stage that are combination
of background (N) and object (RO). As before, the regions will be merged based on merging rule
and the stage will be repeated iteratively and updated and stops when the entire region N can not
find new region for merging. These stages will be repeated again until the merging is not occured.
Finaly, all remaining regions in N will be labeled as object and merged into RO and we can easily
extract the object from the image.
4. EXPERIMENTAL RESULTS
Our results show that the proposed method is very powerfull and robust for object recognition and
detection. We trained three objects (rack, bottle and calendar) from an online dataset [16] in
several views with different scales and illuminations for checking our method. Using ASIFT
keypoints in the region merging algorithm based on maximal similarity, helped us to obtain
significant results. These keypoints are very exact and come from a strong algorithm (ASIFT)
with statistical basis. Moreover, the keypoints give us the best information of objects which are
very useful for merging process.
Figure 7, shows ASIFT keypoints on the images with initial mean shift segmentation and detected
objects with their boundaries. First row, (a), is relevant to the rack dataset, second row, (b), is the
bottle and thirth row, (c), is the calendar.
8. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
36
(a)
(b)
(c)
Figure 7. Left column: ASIFT keypoints with initial mean shift segmentation
right column: detected object with its boundary.
Accuracy rate (6) of full boundary detection for each object is computed and shown in Table 1.
Results show that the accuracy rate of datasets is close together.
do
to
N
N
RateAccuracy ×= 100 (6)
Where Nto indicates the number of trained images of each dataset and Ndo means the number of
full detected objects respective to the dataset.
9. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
37
Table 1. Accuracy rate of three datasets.
Dataset No. of trained
images
No. of tested
images
No. of full
detected object
Accuracy
rate
Rack 60 30 26 86.66 %
Bottle 60 30 28 93.33 %
Calendar 60 30 29 96.66 %
In the previous paper [17], we used SIFT keypoints to recognize the objects and utilized City
Block distance as similarity measure to compare the regions. But, in this paper, we have used
ASIFT keypoints instead of SIFT keypoints. Also, the Euclidean distance has been replaced with
the City Block distance. Comparison between detection rates of [17] and the proposed algorithm
in this paper, shows that the proposed method is more accurate than [17] for object recognition
and detection. For this Comparison, we have applied the proposed algorithm on 3 used dataset in
[17]. The results of the comparison is shown in Table 2.
Table 2. Comparison between detection rates of [17] and the proposed algorithm.
(The rates are relevant to 3 used dataset in [17])
Dataset Detection rate by
method in [17]
Detection rate by
proposed algorithm
Book 91.66 % 95.83 %
Monitor 87.5 % 91.66 %
Stapler 83.33 % 91.66 %
Also, some of the other experimental results are shown in Figure 8.
(a)
10. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
38
(b)
(c)
Figure 8. Left column: one of the trained images for an object, right column:
the detected object. (a): rack, (b): bottle (c): calendar.
5. CONCLUSIONS
This paper presented an object recognition and detection algorithm. These targets are achieved by
combining ASIFT and a region merging segmentation algorithm based on a similarity measure.
We train different objects seperately in several images with multiple aspects and camera
viewpoints to find the best keypoints for recognizing them in the other images. These keypoints
will be applied to the region merging algorithm. The merging process is started by using
keypoints and presented similarity measure (Euclidean distance). The regions will be merged
based on the merging role and Finaly, the object will be detected well, with its boundary. A final
conclusion is that the more keypoints are obtained, and the more accurate they are, the results will
be better and more acceptable. Currently, we are working to develop an approach for shape
recognition of objects in images. We hope to present a robust algorithm for shape recognition by
using the extracted boundaries of objects based on the proposed algorithm in this paper.
11. Signal & Image Processing : An International Journal (SIPIJ) Vol.3, No.5, October 2012
39
REFERENCES
[1] Jean-Michel Morel, Guoshen Yu, (2009) “ASIFT: A New Framework for Fully Affine Invariant
Image Comparison”, SIAM Journal on Imaging Sciences, PP: 438-496.
[2] Guoshen Yu, Jean-Michel Morel, (2009) “A Fully Invariant Image Comparison Method”, IEEE
International Conference on Acoustics, Speech, and Signal Processing (ICASSP), PP: 1597-1600.
[3] David Lowe, (2004) “Distinctive Image Feature Transform from Scale Invariant Keypoints”,
International Journal of Computer Vision, PP: 91-110.
[4] Shivani Agarawal, Aatif Awan, Dan Roth, (2004) “Learning to Detect Objects in Images via a
Sparse, Part-Based Representation”, IEEE Transactions on Pattern Analysis and Machine
Intelligence, PP: 1475-1490.
[5] Constantine Papageorgiou, Tomaso Poggio, (2000) “A Trainable System for Object Detection”.
International Journal of Computer Vision, PP: 15-33.
[6] Dingding Liu, Kari Pulli, Linda G. Shapiro, Yingen Xiong, (2010) “Fast Interactive Image
Segmentation by Discriminative Clustering”, ACM Multimedia Workshop on Mobile Cloud Media
Computing, PP: 47-52.
[7] Bo Peng, Lei Zhang, David Zhang, (2001) “Image Segmentation by Iterated Region Merging with
Localized Graph Cuts”, Pattern Recognition Journal, PP: 2527-2538.
[8] Jifeng Ning, Lei Zhang, David Zhang, Chengke Wu, (2010) “Interactive Image Segmentation by
Maximal Similarity based Region Merging”, Pattern Recognition Journal, PP: 445-456.
[9] David G. Lowe, (1999) ”Object Recognition from Scale Invariant Features”, International conference
of Computer Vision, PP: 1150-1157.
[10] Mathew Brown, David Lowe, (2002) “Invariant Features from Interest Point Groups”, British
Machine Vision Conference, PP: 556-565.
[11] Kristian Mikolajczyk, Cordelia Schmid, (2005) “A Performance Evaluation of Local Descriptor”,
IEEE Transaction on Pattern Analysis and Machine Intelligence, PP: 1615-1630.
[12] Dorin Comaniciu, Peter Meer, (2002) “Mean Shift: A Robust Approach toward Space Analysis”,
IEEE Transactions on Pattern Analysis and Machine Intelligence, PP: 603-619.
[13] EDISON system. (http://coewww.rutgers.edu/riul/research/code/EDISON/index.html).
[14] Jens Kaftan, Andre Bell, Til Aach, (2008) “Mean Shift Segmentation - Evaluation of Optimization
Techniques”, International Conference on Computer Vision Theory and Applications, PP: 365-374.
[15] Qiming Luo, Taghi Khoshgoftaar, (2004) “Efficient Image Segmentation by Mean Shift Clustering
and MDL-Guided Region Merging”, International Conference on Tools with Artificial Intelligence,
PP: 337-343.
[16] Online Dataset. (http://www.iiia.csic.es/~aramisa/datasets/iiia30.html).
[17] Reza Oji, Farshad Tajeripour, (2012) “Full Object Boundary Detection by Applying Scale Invariant
Features in a Region Merging Segmentation Algorithm”, International Journal of Artificial
Intelligence and Applications, PP: 41-50.
Author
Reza oji received the B.Sc. degree in Computer Software Engineering
from Azad University of Shiraz. He received the M.Sc. degree in Artificial
Intelligence from International University of Shiraz. His main research
interests include image processing and machine vision.