This document summarizes a capstone project that developed a vision-based system for classifying turf and non-turf regions for an autonomous lawn mower. The system extracts color and texture features from blocks of the camera image and then classifies each block as turf or non-turf using k-means clustering or support vector machines (SVM). An SVM classifier with a radial basis function kernel produced the best performance across varied lighting conditions. The project aims to allow the mower to avoid non-turf regions and obstacles in real-time using computer vision instead of boundary wires or contact sensors.
The development of multimedia system technology in Content based Image Retrieval (CBIR) System is
one in every of the outstanding area to retrieve the images from an oversized collection of database. The feature
vectors of the query image are compared with feature vectors of the database images to get matching images.It is
much observed that anyone algorithm isn't beneficial in extracting all differing kinds of natural images. Thus an
intensive analysis of certain color, texture and shape extraction techniques are allotted to spot an efficient CBIR
technique that suits for a selected sort of images. The Extraction of an image includes feature description and
feature extraction. During this paper, we tend to projected Color Layout Descriptor (CLD), grey Level Co-
Occurrences Matrix (GLCM), Marker-Controlled Watershed Segmentation feature extraction technique that
extract the matching image based on the similarity of Color, Texture and shape within the database. For
performance analysis, the image retrieval timing results of the projected technique is calculated and compared
with every of the individual feature.
Automated Colorization of Grayscale Images Using Texture DescriptorsIDES Editor
The document proposes a novel automated process called ACTD for colorizing grayscale images using texture descriptors without human intervention. It analyzes sample color images to extract coherent texture regions, then uses Gabor filtering for texture-based segmentation. Texture descriptors and color information are computed and stored for each region. These are then used to colorize a new grayscale image based on texture matching. The method combines techniques such as Gabor filtering, fuzzy C-means clustering with a new "Gki factor" for noise tolerance, and content-based image retrieval of texture descriptors. Preliminary results found the approach viable but improvements were needed for scale and rotation invariance.
A version of watershed algorithm for color image segmentationHabibur Rahman
The document summarizes a master's thesis presentation on a new watershed algorithm for color image segmentation. The thesis addresses issues with existing watershed algorithms like over-segmentation and sensitivity to noise. The contributions of the thesis include an adaptive masking and thresholding mechanism to overcome over-segmentation and perform well on noisy images. The thesis is evaluated using five image quality assessment metrics on 20 classes of images, showing the proposed method performs better and has lower computational complexity than other algorithms. In conclusions, the adaptive watershed algorithm ensures accurate segmentation and is suitable for real-time applications.
1. The document presents an image segmentation algorithm that uses local thresholding in the YCbCr color space.
2. It computes local thresholds for each pixel by calculating the mean and standard deviation of neighboring pixels in a 3x3 mask. The threshold is used to label each pixel as 1 or 0.
3. The algorithm was tested on images with objects indistinct and distinct from the background. It performed well in segmenting objects from the background in both cases. There is potential to improve performance for blurred images.
The document discusses implementing the watershed algorithm for image segmentation using MATLAB. It begins with an overview of the watershed algorithm which uses region growing to segment images based on gradients. It then discusses image segmentation goals and different categories of segmentation algorithms including discontinuity, similarity, thresholding and region-based approaches. The document outlines steps to be taken which include preprocessing images, implementing the watershed algorithm, and analyzing results to see the difference in segmented images. It expects watershed processing to produce segmented regions from a test image.
Image segmentation involves grouping similar image components, such as pixels, into segments. It has applications in medical imaging, satellite imagery, and video summarization. Common methods include thresholding, k-means clustering, and region-based approaches. Thresholding segments an image based on pixel intensity values, while k-means clustering groups pixels into a specified number of clusters based on color or other feature similarity. Region-based methods grow or merge regions of similar pixels. Watershed segmentation treats an image as a topographic surface and finds boundaries between regions.
1. Image classification involves categorizing all pixels in a remote sensing image based on their spectral patterns. There are two main approaches: supervised classification, where an operator identifies training areas, and unsupervised classification, where software automatically groups pixels into clusters.
2. Accuracy assessment is done by comparing classified data to reference data using an error matrix. The matrix shows errors of omission and commission to calculate overall, user's, and producer's accuracy percentages. The kappa coefficient provides an overall measure of classification accuracy.
3. Classified images undergo post-processing like filtering to improve accuracy before being used as thematic maps, tables, or GIS inputs. Problems in urban areas include similar surface spectra and mixed pixel issues.
The development of multimedia system technology in Content based Image Retrieval (CBIR) System is
one in every of the outstanding area to retrieve the images from an oversized collection of database. The feature
vectors of the query image are compared with feature vectors of the database images to get matching images.It is
much observed that anyone algorithm isn't beneficial in extracting all differing kinds of natural images. Thus an
intensive analysis of certain color, texture and shape extraction techniques are allotted to spot an efficient CBIR
technique that suits for a selected sort of images. The Extraction of an image includes feature description and
feature extraction. During this paper, we tend to projected Color Layout Descriptor (CLD), grey Level Co-
Occurrences Matrix (GLCM), Marker-Controlled Watershed Segmentation feature extraction technique that
extract the matching image based on the similarity of Color, Texture and shape within the database. For
performance analysis, the image retrieval timing results of the projected technique is calculated and compared
with every of the individual feature.
Automated Colorization of Grayscale Images Using Texture DescriptorsIDES Editor
The document proposes a novel automated process called ACTD for colorizing grayscale images using texture descriptors without human intervention. It analyzes sample color images to extract coherent texture regions, then uses Gabor filtering for texture-based segmentation. Texture descriptors and color information are computed and stored for each region. These are then used to colorize a new grayscale image based on texture matching. The method combines techniques such as Gabor filtering, fuzzy C-means clustering with a new "Gki factor" for noise tolerance, and content-based image retrieval of texture descriptors. Preliminary results found the approach viable but improvements were needed for scale and rotation invariance.
A version of watershed algorithm for color image segmentationHabibur Rahman
The document summarizes a master's thesis presentation on a new watershed algorithm for color image segmentation. The thesis addresses issues with existing watershed algorithms like over-segmentation and sensitivity to noise. The contributions of the thesis include an adaptive masking and thresholding mechanism to overcome over-segmentation and perform well on noisy images. The thesis is evaluated using five image quality assessment metrics on 20 classes of images, showing the proposed method performs better and has lower computational complexity than other algorithms. In conclusions, the adaptive watershed algorithm ensures accurate segmentation and is suitable for real-time applications.
1. The document presents an image segmentation algorithm that uses local thresholding in the YCbCr color space.
2. It computes local thresholds for each pixel by calculating the mean and standard deviation of neighboring pixels in a 3x3 mask. The threshold is used to label each pixel as 1 or 0.
3. The algorithm was tested on images with objects indistinct and distinct from the background. It performed well in segmenting objects from the background in both cases. There is potential to improve performance for blurred images.
The document discusses implementing the watershed algorithm for image segmentation using MATLAB. It begins with an overview of the watershed algorithm which uses region growing to segment images based on gradients. It then discusses image segmentation goals and different categories of segmentation algorithms including discontinuity, similarity, thresholding and region-based approaches. The document outlines steps to be taken which include preprocessing images, implementing the watershed algorithm, and analyzing results to see the difference in segmented images. It expects watershed processing to produce segmented regions from a test image.
Image segmentation involves grouping similar image components, such as pixels, into segments. It has applications in medical imaging, satellite imagery, and video summarization. Common methods include thresholding, k-means clustering, and region-based approaches. Thresholding segments an image based on pixel intensity values, while k-means clustering groups pixels into a specified number of clusters based on color or other feature similarity. Region-based methods grow or merge regions of similar pixels. Watershed segmentation treats an image as a topographic surface and finds boundaries between regions.
1. Image classification involves categorizing all pixels in a remote sensing image based on their spectral patterns. There are two main approaches: supervised classification, where an operator identifies training areas, and unsupervised classification, where software automatically groups pixels into clusters.
2. Accuracy assessment is done by comparing classified data to reference data using an error matrix. The matrix shows errors of omission and commission to calculate overall, user's, and producer's accuracy percentages. The kappa coefficient provides an overall measure of classification accuracy.
3. Classified images undergo post-processing like filtering to improve accuracy before being used as thematic maps, tables, or GIS inputs. Problems in urban areas include similar surface spectra and mixed pixel issues.
A DIGITAL COLOR IMAGE WATERMARKING SYSTEM USING BLIND SOURCE SEPARATIONcsandit
An attempt is made to implement a digital color image-adaptive watermarking scheme in
spatial domain and hybrid domain i.e host image in wavelet domain and watermark in spatial
domain. Blind Source Separation (BSS) is used to extract the watermark The novelty of the
presented scheme lies in determining the mixing matrix for BSS model using BFGS (Broyden–
Fletcher–Goldfarb–Shanno) optimization technique. This method is based on the smooth and
textured portions of the image. Texture analysis is carried based on energy content of the
image (using GLCM) which makes the method image adaptive to embed color watermark.
The performance evaluation is carried for hybrid domain of various color spaces like YIQ, HSI
and YCbCr and the feasibility of optimization algorithm for finding mixing matrix is also
checked for these color spaces. Three ICA (Independent Component Analysis)/BSS algorithms
are used in extraction procedure ,through which the watermark can be retrieved efficiently . An
effort is taken to find out the best suited color space to embed the watermark which satisfies the
condition of imperceptibility and robustness against various attacks.
An evaluation of two popular segmentation algorithms, the mean shift-based segmentation algorithm and a graph-based segmentation scheme. We also consider a hybrid method which combines the other two methods.
Object detection for service robot using range and color features of an imageIJCSEA Journal
In real-world applications, service robots need to locate and identify objects in a scene. A range sensor
provides a robust estimate of depth information, which is useful to accurately locate objects in a scene. On
the other hand, color information is an important property for object recognition task. The objective of this
paper is to detect and localize multiple objects within an image using both range and color features. The
proposed method uses 3D shape features to generate promising hypotheses within range images and
verifies these hypotheses by using features obtained from both range and color images.
Object Detection for Service Robot Using Range and Color Features of an ImageIJCSEA Journal
In real-world applications, service robots need to locate and identify objects in a scene. A range sensor provides a robust estimate of depth information, which is useful to accurately locate objects in a scene. On the other hand, color information is an important property for object recognition task. The objective of this paper is to detect and localize multiple objects within an image using both range and color features. The proposed method uses 3D shape features to generate promising hypotheses within range images and verifies these hypotheses by using features obtained from both range and color images.
EFFICIENT IMAGE RETRIEVAL USING REGION BASED IMAGE RETRIEVALsipij
1) The document describes an efficient region-based image retrieval system that uses discrete wavelet transform and k-means clustering. It segments images into regions, each characterized by features like size, mean, and covariance.
2) The system pre-processes images by resizing, converting to HSV color space, performing DWT, and using k-means clustering on DWT coefficients to generate regions. It extracts features for each region and stores them in a database.
3) For retrieval, it pre-processes the query image similarly and calculates similarities between the query regions and database regions based on their features, returning similar images.
This document summarizes a research paper on background subtraction techniques for motion detection in video. It describes a proposed technique that stores and compares past pixel values to the current value to determine if a pixel belongs to the background or foreground. It also discusses using a k-means algorithm and Gaussian mixture model to build a probabilistic background model and classify pixels. The paper evaluates different shadow detection approaches and finds RGB color spaces perform best for segmentation and shadow removal.
IMAGE SEGMENTATION BY USING THRESHOLDING TECHNIQUES FOR MEDICAL IMAGEScseij
Image binarization is the process of separation of pixel values into two groups, black as background and
white as foreground. Thresholding can be categorized into global thresholding and local thresholding. This
paper describes a locally adaptive thresholding technique that removes background by using local mean
and standard deviation. Most common and simplest approach to segment an image is using thresholding.
In this work we present an efficient implementation for threshoding and give a detailed comparison of
Niblack and sauvola local thresholding algorithm. Niblack and sauvola thresholding algorithm is
implemented on medical images. The quality of segmented image is measured by statistical parameters:
Jaccard Similarity Coefficient, Peak Signal to Noise Ratio (PSNR).
Marker Controlled Segmentation Technique for Medical applicationRushin Shah
Medical image segmentation is a very important field for the medical science. In medical images, edge detection is an important work for object recognition of the human organs such as brain, heart or kidney etc. and it is an essential pre-processing step in medical image segmentation.
Medical images such as CT, MRI or X-Ray visualizes the various information’s of internal organs which is very important for doctors diagnoses as well as medical teaching, learning and research.
It is a tough job to locate the internal organs if images contains noise or rough structure of human body organs.
The document discusses image segmentation techniques. It describes image segmentation as partitioning a digital image into multiple regions based on characteristics like color or texture. Common applications of image segmentation include industrial inspection, optical character recognition, and medical imaging. The techniques discussed are fixed thresholding, iterative thresholding, and fuzzy c-means clustering. Fuzzy c-means clustering is identified as the most suitable for pest image segmentation based on its lower entropy and normalized mutual information values. Simulated annealing is also proposed to improve upon the limitations of fuzzy c-means clustering.
Digital image classification involves:
1) Sorting pixels into classes based on their spectral values using algorithms like supervised maximum likelihood classification or unsupervised isodata clustering.
2) Analyzing spectral patterns by examining pixels in feature space rather than image space. Distances between pixel vectors in feature spaces define class boundaries.
3) Validating classification results to determine accuracy by comparing to reference data. Problems can occur and techniques continue improving.
NMS and Thresholding Architecture used for FPGA based Canny Edge Detector for...idescitation
In this paper, an architecture designed for Non-
Maximal Suppression used in Canny edge detection algorithm
is presented in order to reduce memory requirements
significantly. The architecture also achieves decreased latency
and increased throughput with no loss in edge detection. The
new algorithm used has a low-complexity 8-bin non-uniform
gradient magnitude histogram to compute block-based
hysteresis thresholds that are used by the Canny edge detector.
Furthermore, the hardware architecture of the proposed
algorithm is presented in this paper and the architecture is
synthesized on the Xilinx Virtex 5 FPGA. The design
development is done in VHDL and simulated results are
obtained using modelsim 6.3 with Xilinx 12.2.
At the end of this lesson, you should be able to;
describe Connected Components and Contours in image segmentation.
discuss region based segmentation method.
discuss Region Growing segmentation technique.
discuss Morphological Watersheds segmentation.
discuss Model Based Segmentation.
discuss Motion Segmentation.
implement connected components, flood fill, watershed, template matching and frame difference techniques.
formulate possible mechanisms to propose segmentation methods to solve problems.
The mean shift procedure is a general nonparametric technique for analyzing complex multimodal feature spaces and delineating arbitrarily shaped clusters. It works by recursively finding the nearest stationary point of the underlying density function, which corresponds to the mode of the density. The mean shift procedure relates to kernel density estimation and robust M-estimators of location. It provides a versatile tool for feature space analysis that can solve many low-level computer vision tasks with few parameters.
Importance of Mean Shift in Remote Sensing SegmentationIOSR Journals
1) Mean shift is a non-parametric clustering technique that can segment remote sensing images into homogeneous regions without prior knowledge of the number of clusters or constraints on cluster shape.
2) The document presents a case study demonstrating mean shift can segment an image containing oil storage tanks into distinct regions faster than level set segmentation.
3) Mean shift is shown to be well-suited for remote sensing image segmentation tasks like forest mapping and land cover classification due to its ability to handle noise, gradients, and texture variations common in real-world images.
A Novel Background Subtraction Algorithm for Dynamic Texture ScenesIJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
This document summarizes a research paper that proposes a new technique for binarizing images captured of black/green boards using a mobile camera. It begins with an abstract that overviews binarizing degraded images from mobile-captured black/green board images to extract text with 92.589% accuracy. It then reviews existing binarization techniques in the literature and describes common global and local thresholding methods. The proposed technique enhances the input image, segments it into 3x3 parts, computes local thresholds using OTSU for each part, binarizes the parts, and joins them. Experimental results on a database of 50 mobile-captured board images show the technique achieves better accuracy than other algorithms according to evaluation metrics.
This document summarizes a research paper that presents a real-time 3D reconstruction method using stereo vision from a driving car. The method extends LSD-SLAM with stereo capabilities to simultaneously track camera pose and reconstruct semi-dense depth maps. It is evaluated on the KITTI dataset and compared to laser scans and traditional stereo methods. Results show the direct SLAM technique generates visually pleasing and globally consistent semi-dense reconstructions in real-time on a single CPU.
At the end of this lesson, you should be able to;
define segmentation.
Describe edge based in segmentation.
describe thresholding and its properties.
apply edge detection and thresholding as segmentation techniques.
etailment WIEN 2016 – Jochen Felsberger – SmartInfoBroker – Beacons in der Pr...Werbeplanung.at Summit
Beacons als neueste Technologie für Location Based Services sind zurzeit in aller Munde. Aber was genau tun sie und was kann man mit ihnen tun? Die Antworten darauf sind für jede Wirtschaftssparte verschieden. In diesem Workshop haben Sie die Möglichkeit, mit einem Hersteller von Beacon-basierten LBS-Lösungen über das Potential solcher Lösungen für Ihre spezifische Branche zu diskutieren, und dabei einige mögliche Anwendungen anhand eines realen Test-Systems live mitzuerleben.
Visualr is a Data Visualization Tool. It's been designed to ease the overall process of Dashboarding. Dashboards are basically required to help in the managerial decision making.
The document summarizes a meeting between a client and design/marketing team to develop a joint strategy. They agreed on the target audience and key messaging aspects like tone, form and color. They planned content for blogs, Facebook, Instagram and analyzed data to build an engaging online community. Their strategy included email marketing, Facebook ads, Google ads, YouTube and Pinterest to engage the audience and gain loyalty.
A DIGITAL COLOR IMAGE WATERMARKING SYSTEM USING BLIND SOURCE SEPARATIONcsandit
An attempt is made to implement a digital color image-adaptive watermarking scheme in
spatial domain and hybrid domain i.e host image in wavelet domain and watermark in spatial
domain. Blind Source Separation (BSS) is used to extract the watermark The novelty of the
presented scheme lies in determining the mixing matrix for BSS model using BFGS (Broyden–
Fletcher–Goldfarb–Shanno) optimization technique. This method is based on the smooth and
textured portions of the image. Texture analysis is carried based on energy content of the
image (using GLCM) which makes the method image adaptive to embed color watermark.
The performance evaluation is carried for hybrid domain of various color spaces like YIQ, HSI
and YCbCr and the feasibility of optimization algorithm for finding mixing matrix is also
checked for these color spaces. Three ICA (Independent Component Analysis)/BSS algorithms
are used in extraction procedure ,through which the watermark can be retrieved efficiently . An
effort is taken to find out the best suited color space to embed the watermark which satisfies the
condition of imperceptibility and robustness against various attacks.
An evaluation of two popular segmentation algorithms, the mean shift-based segmentation algorithm and a graph-based segmentation scheme. We also consider a hybrid method which combines the other two methods.
Object detection for service robot using range and color features of an imageIJCSEA Journal
In real-world applications, service robots need to locate and identify objects in a scene. A range sensor
provides a robust estimate of depth information, which is useful to accurately locate objects in a scene. On
the other hand, color information is an important property for object recognition task. The objective of this
paper is to detect and localize multiple objects within an image using both range and color features. The
proposed method uses 3D shape features to generate promising hypotheses within range images and
verifies these hypotheses by using features obtained from both range and color images.
Object Detection for Service Robot Using Range and Color Features of an ImageIJCSEA Journal
In real-world applications, service robots need to locate and identify objects in a scene. A range sensor provides a robust estimate of depth information, which is useful to accurately locate objects in a scene. On the other hand, color information is an important property for object recognition task. The objective of this paper is to detect and localize multiple objects within an image using both range and color features. The proposed method uses 3D shape features to generate promising hypotheses within range images and verifies these hypotheses by using features obtained from both range and color images.
EFFICIENT IMAGE RETRIEVAL USING REGION BASED IMAGE RETRIEVALsipij
1) The document describes an efficient region-based image retrieval system that uses discrete wavelet transform and k-means clustering. It segments images into regions, each characterized by features like size, mean, and covariance.
2) The system pre-processes images by resizing, converting to HSV color space, performing DWT, and using k-means clustering on DWT coefficients to generate regions. It extracts features for each region and stores them in a database.
3) For retrieval, it pre-processes the query image similarly and calculates similarities between the query regions and database regions based on their features, returning similar images.
This document summarizes a research paper on background subtraction techniques for motion detection in video. It describes a proposed technique that stores and compares past pixel values to the current value to determine if a pixel belongs to the background or foreground. It also discusses using a k-means algorithm and Gaussian mixture model to build a probabilistic background model and classify pixels. The paper evaluates different shadow detection approaches and finds RGB color spaces perform best for segmentation and shadow removal.
IMAGE SEGMENTATION BY USING THRESHOLDING TECHNIQUES FOR MEDICAL IMAGEScseij
Image binarization is the process of separation of pixel values into two groups, black as background and
white as foreground. Thresholding can be categorized into global thresholding and local thresholding. This
paper describes a locally adaptive thresholding technique that removes background by using local mean
and standard deviation. Most common and simplest approach to segment an image is using thresholding.
In this work we present an efficient implementation for threshoding and give a detailed comparison of
Niblack and sauvola local thresholding algorithm. Niblack and sauvola thresholding algorithm is
implemented on medical images. The quality of segmented image is measured by statistical parameters:
Jaccard Similarity Coefficient, Peak Signal to Noise Ratio (PSNR).
Marker Controlled Segmentation Technique for Medical applicationRushin Shah
Medical image segmentation is a very important field for the medical science. In medical images, edge detection is an important work for object recognition of the human organs such as brain, heart or kidney etc. and it is an essential pre-processing step in medical image segmentation.
Medical images such as CT, MRI or X-Ray visualizes the various information’s of internal organs which is very important for doctors diagnoses as well as medical teaching, learning and research.
It is a tough job to locate the internal organs if images contains noise or rough structure of human body organs.
The document discusses image segmentation techniques. It describes image segmentation as partitioning a digital image into multiple regions based on characteristics like color or texture. Common applications of image segmentation include industrial inspection, optical character recognition, and medical imaging. The techniques discussed are fixed thresholding, iterative thresholding, and fuzzy c-means clustering. Fuzzy c-means clustering is identified as the most suitable for pest image segmentation based on its lower entropy and normalized mutual information values. Simulated annealing is also proposed to improve upon the limitations of fuzzy c-means clustering.
Digital image classification involves:
1) Sorting pixels into classes based on their spectral values using algorithms like supervised maximum likelihood classification or unsupervised isodata clustering.
2) Analyzing spectral patterns by examining pixels in feature space rather than image space. Distances between pixel vectors in feature spaces define class boundaries.
3) Validating classification results to determine accuracy by comparing to reference data. Problems can occur and techniques continue improving.
NMS and Thresholding Architecture used for FPGA based Canny Edge Detector for...idescitation
In this paper, an architecture designed for Non-
Maximal Suppression used in Canny edge detection algorithm
is presented in order to reduce memory requirements
significantly. The architecture also achieves decreased latency
and increased throughput with no loss in edge detection. The
new algorithm used has a low-complexity 8-bin non-uniform
gradient magnitude histogram to compute block-based
hysteresis thresholds that are used by the Canny edge detector.
Furthermore, the hardware architecture of the proposed
algorithm is presented in this paper and the architecture is
synthesized on the Xilinx Virtex 5 FPGA. The design
development is done in VHDL and simulated results are
obtained using modelsim 6.3 with Xilinx 12.2.
At the end of this lesson, you should be able to;
describe Connected Components and Contours in image segmentation.
discuss region based segmentation method.
discuss Region Growing segmentation technique.
discuss Morphological Watersheds segmentation.
discuss Model Based Segmentation.
discuss Motion Segmentation.
implement connected components, flood fill, watershed, template matching and frame difference techniques.
formulate possible mechanisms to propose segmentation methods to solve problems.
The mean shift procedure is a general nonparametric technique for analyzing complex multimodal feature spaces and delineating arbitrarily shaped clusters. It works by recursively finding the nearest stationary point of the underlying density function, which corresponds to the mode of the density. The mean shift procedure relates to kernel density estimation and robust M-estimators of location. It provides a versatile tool for feature space analysis that can solve many low-level computer vision tasks with few parameters.
Importance of Mean Shift in Remote Sensing SegmentationIOSR Journals
1) Mean shift is a non-parametric clustering technique that can segment remote sensing images into homogeneous regions without prior knowledge of the number of clusters or constraints on cluster shape.
2) The document presents a case study demonstrating mean shift can segment an image containing oil storage tanks into distinct regions faster than level set segmentation.
3) Mean shift is shown to be well-suited for remote sensing image segmentation tasks like forest mapping and land cover classification due to its ability to handle noise, gradients, and texture variations common in real-world images.
A Novel Background Subtraction Algorithm for Dynamic Texture ScenesIJMER
International Journal of Modern Engineering Research (IJMER) is Peer reviewed, online Journal. It serves as an international archival forum of scholarly research related to engineering and science education.
This document summarizes a research paper that proposes a new technique for binarizing images captured of black/green boards using a mobile camera. It begins with an abstract that overviews binarizing degraded images from mobile-captured black/green board images to extract text with 92.589% accuracy. It then reviews existing binarization techniques in the literature and describes common global and local thresholding methods. The proposed technique enhances the input image, segments it into 3x3 parts, computes local thresholds using OTSU for each part, binarizes the parts, and joins them. Experimental results on a database of 50 mobile-captured board images show the technique achieves better accuracy than other algorithms according to evaluation metrics.
This document summarizes a research paper that presents a real-time 3D reconstruction method using stereo vision from a driving car. The method extends LSD-SLAM with stereo capabilities to simultaneously track camera pose and reconstruct semi-dense depth maps. It is evaluated on the KITTI dataset and compared to laser scans and traditional stereo methods. Results show the direct SLAM technique generates visually pleasing and globally consistent semi-dense reconstructions in real-time on a single CPU.
At the end of this lesson, you should be able to;
define segmentation.
Describe edge based in segmentation.
describe thresholding and its properties.
apply edge detection and thresholding as segmentation techniques.
etailment WIEN 2016 – Jochen Felsberger – SmartInfoBroker – Beacons in der Pr...Werbeplanung.at Summit
Beacons als neueste Technologie für Location Based Services sind zurzeit in aller Munde. Aber was genau tun sie und was kann man mit ihnen tun? Die Antworten darauf sind für jede Wirtschaftssparte verschieden. In diesem Workshop haben Sie die Möglichkeit, mit einem Hersteller von Beacon-basierten LBS-Lösungen über das Potential solcher Lösungen für Ihre spezifische Branche zu diskutieren, und dabei einige mögliche Anwendungen anhand eines realen Test-Systems live mitzuerleben.
Visualr is a Data Visualization Tool. It's been designed to ease the overall process of Dashboarding. Dashboards are basically required to help in the managerial decision making.
The document summarizes a meeting between a client and design/marketing team to develop a joint strategy. They agreed on the target audience and key messaging aspects like tone, form and color. They planned content for blogs, Facebook, Instagram and analyzed data to build an engaging online community. Their strategy included email marketing, Facebook ads, Google ads, YouTube and Pinterest to engage the audience and gain loyalty.
The document outlines updates from various Bonner Foundation meetings, programs, and initiatives. Key points include:
1) Upcoming meetings include the IMPACT Conference in February and the Summer Leadership Institute in May.
2) The Bonner Congress theme is "Ideas to Action" with an emphasis on communication between campuses and annual reporting.
3) Foundation support includes staff campus visits, wiki improvements, and potential fundraising resources. National partner opportunities are highlighted weekly.
4) Areas of focus include capstone projects, community-based research, faculty development, student engagement, food security, and college access. Assessment studies include impact data and financial aid surveys.
Día internacional de la paz, Naciones Unidas - 21 septiembre - ¿Qués es?Unair Cast
Es el día que la Asamblea General de las Naciones Unidas aprobó por unanimidad la resolución 55/282 , que estableció el 21 de septiembre como un día de cesación del fuego y de no violencia a nivel mundial.
Así cada 21 de septiembre se celebra el Día Internacional de la Paz.
en la Constitución de la Organización de las Naciones Unidas para la Educación, la Ciencia y la Cultura se declara que “puesto que las guerras nacen en la mente de los hombres, es en la mente de los hombres donde deben erigirse los baluartes de la paz”
Expresando profunda preocupación por la persistencia y la proliferación de la violencia y los conflictos en diversas partes del mundo, Reconociendo la necesidad de eliminar todas las formas de discriminación e intolerancia, incluidas las basadas en la raza, el color, el sexo, el idioma, la religión, la opinión política o de otra índole, el origen nacional, étnico o social, la propiedad, las discapacidades, el nacimiento u otra condición
La paz no sólo es la ausencia de conflictos, sino que también requiere un proceso positivo, dinámico y participativo en que se promueva el diálogo y se solucionen los conflictos en un espíritu de entendimiento y cooperación mutuos
La ONU proclama solemnemente la Declaración sobre una Cultura de Paz con el fin de que los Gobiernos, las organizaciones internacionales y la sociedad civil puedan orientar sus actividades por sus disposiciones a fin de promover y fortalecer una cultura de paz en el nuevo milenio
Y se invita a todas las naciones y pueblos a que cumplan una cesación de hostilidades durante todo ese Día y a que también lo celebren mediante la educación y la sensibilización del público sobre todos los temas relacionados con la paz.
سلام
Ειρήνη
Frieden
平和
평화
Мир
שלום
Darwin's theory of evolution by natural selection proposes that organisms gradually change over generations through a "descent with modification" process where favorable inherited traits become more common in a population. Evidence from fields such as paleontology, genetics and developmental biology support this theory. Scientists have found fossils that show whales evolved from land mammals, starting as four-legged creatures that walked both on land and swam in water, gradually becoming fully aquatic and losing legs over millions of years. While controversial for some, evolution is well-established scientific fact that explains the diversity of life.
Why is Open Innovation important? What lessons can we learn from case studies? How should companies, both startups and large corporates, actually implement Open Innovation?
The Met and Museum Collections on TwitterNeal Stimler
This slide deck is associated with the talk at Cité des sciences et de l'industrie, Paris hosted by CLIC France on January 15, 2016. Download for all active links.
Werbeplanung.at SUMMIT 16 – New Marketing on New Platforms – Coca-Coal Case S...Werbeplanung.at Summit
Was für welchen Zweck und welche passt zu meinem Unternehmen? Zielgruppen, Content, die Motivation, Inhalte zu teilen: Von Plattform zu Plattform gibt es da gewaltige Unterschiede und daraus ergeben sich ganz unterschiedliche Implikationen. Jasmin Fischer Social-Media-Profi von Coca-Cola zeigte am Werbeplanung.at SUMMIT mit Coca-Cola ein erfolgreiches Beispiel.
NEW AGE FINANCIAL PRODUCTS & THEIR ROLE IN TRANSFORMING MODERN BANKINGRADHIKA GUPTA
This document provides a summary of a student paper titled "NEW AGE FINANCIAL PRODUCTS & THEIR ROLE IN TRANSFORMING MODERN BANKING". The paper examines how new banking products like debit cards, credit cards, POS machines, ATMs, e-banking and SMS banking have transformed modern banking. It includes a literature review on past studies and sections analyzing each financial product and how they have changed customer behaviors and transactions. The conclusion finds that these new products have significantly changed buying and transaction habits and banking scenarios in recent times.
This document provides details about typical cross-sections of roads and highways, including pavement surfaces and drainage elements. It discusses the importance of friction between wheels and pavement, pavement smoothness, light reflection characteristics, and drainage. It also describes typical layers in flexible pavements like seal coats, surface courses, binder courses, and subgrades. Finally, it outlines other cross-section elements such as shoulders, medians, footpaths, barriers, and bus bays.
Apple's 2015 financial results showed improved profitability, efficiency, and liquidity compared to the previous period. Key metrics like gross profit margin, net profit margin, return on assets and return on investment all increased. Asset turnover ratio and stock turnover ratio improved, demonstrating higher efficiency. Liquidity ratios such as working capital ratio, quick asset ratio and cash flow ratio were also better. However, debt ratio increased, indicating lower stability as more assets were financed through liabilities.
A STUDY ON INTERNET BANKING SERVICE QUALITY IN CACHAR DISTRICTSandip Dey
This document is a dissertation submitted by Sandip Dey for the partial fulfillment of a Master of Business Administration degree. The dissertation is titled "A Study on Internet Banking Service Quality in Cachar District". It includes an introduction, acknowledgements, preface, table of contents, and the first chapter which provides an overview of internet banking, measuring service quality, parameters for measuring service quality, capturing customer perspectives, gaps in service design and delivery, and strategies for closing gaps. The document appears to be analyzing service quality of internet banking in Cachar District in India.
Unsupervised Building Extraction from High Resolution Satellite Images Irresp...CSCJournals
The document discusses an unsupervised method for extracting buildings from high resolution satellite images regardless of rooftop structures. The method first calculates NDVI and chromaticity ratios to segment vegetation and shadows. Rooftops and roads are then detected and eliminated. Principal component analysis and area analysis are performed to accurately extract buildings. The algorithm aims to eliminate inhomogeneities caused by varying building hierarchies by focusing on eliminating non-building regions rather than detecting building regions of interest. The methodology is tested on Quickbird satellite imagery and results indicate it can extract buildings in complex environments irrespective of rooftop shape.
1) The document describes a method for vehicle detection and tracking in videos using inter-frame coding. It extracts moving blobs through background subtraction and clusters pixels into regions.
2) A blob merging module merges close-by blobs that were incorrectly separated due to occlusion. Tracking corresponds blobs between frames using position, color, and velocity.
3) Experimental results on real-time video show the method can successfully detect and track vehicles under different lighting conditions and levels of noise. The addition of a "Vicinity Factor" improved segmentation accuracy compared to using blob merging thresholds alone.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Object Classification of Satellite Images Using Cluster Repulsion Based Kerne...IOSR Journals
Abstract: We investigated the Classification of satellite images and multispectral remote sensing data .we
focused on uncertainty analysis in the produced land-cover maps .we proposed an efficient technique for
classifying the multispectral satellite images using Support Vector Machine (SVM) into road area, building area
and green area. We carried out classification in three modules namely (a) Preprocessing using Gaussian
filtering and conversion from conversion of RGB to Lab color space image (b) object segmentation using
proposed Cluster repulsion based kernel Fuzzy C- Means (FCM) and (c) classification using one-to-many SVM
classifier. The goal of this research is to provide the efficiency in classification of satellite images using the
object-based image analysis. The proposed work is evaluated using the satellite images and the accuracy of the
proposed work is compared to FCM based classification. The results showed that the proposed technique has
achieved better results reaching an accuracy of 79%, 84%, 81% and 97.9% for road, tree, building and vehicle
classification respectively.
Keywords:-Satellite image, FCM Clustering, Classification, SVM classifier.
This document discusses using convolutional neural networks (CNNs) to classify and segment satellite imagery. It presents a novel approach using a CNN to perform per-pixel classification of multispectral satellite imagery and a digital surface model into five categories (vegetation, ground, roads, buildings, water). The CNN is first pre-trained with unsupervised clustering then fine-tuned for classification and segmentation. Results show the CNN approach outperforms existing methods, achieving 94.49% classification accuracy and improving segmentation by reducing salt-and-pepper effects from per-pixel classification alone.
OBJECT DETECTION FOR SERVICE ROBOT USING RANGE AND COLOR FEATURES OF AN IMAGEIJCSEA Journal
This document summarizes an approach for object detection using both range and color image features. The proposed method first generates hypotheses for objects in a range image using a generative model (pLSA) applied to bag-of-visual-words representing 3D shape. It then verifies the hypotheses using an SVM classifier combining 3D shape features from the range image and color appearance features from the corresponding area of the color image. The approach was tested on images containing multiple objects acquired using both a range sensor and color camera.
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
This document summarizes research comparing three machine learning classification methods - Decision Tree, Support Vector Machine (SVM), and k-Nearest Neighbors (k-NN) - for classifying land use from high and low resolution satellite imagery. The researchers applied each method to classify Pleiades satellite images of Taiwan and Colorado. SVM achieved the highest overall accuracy of 78.6% for high resolution imagery and 83.3% for low resolution imagery. Decision Trees and k-NN were less accurate. The document outlines the methodology, including image preprocessing, parameter selection, accuracy assessment, and findings.
Detection of urban tree canopy from very high resolution imagery using an ob...IJECEIAES
Tree that grows within a town, city and suburban areas, collection of these trees makes the urban forest. These urban forest and urban trees have impact on urban water, pollution and heat. Nowadays we are experiencing drastic climatic changes because of cutting of trees for our growth and increasing population which leads to expansion of roads, towers, and airports. Individual tree crown detection is necessary to map the forest along with feasible planning for urban areas. In this study, using WorldView-2imagery, trees in specific area are detected with object-based image analysis (OBAI) approach. Therefore with improvement in spatial and spectral resolution of an image, extracted from WorldView-2 carried out urban features with better accuracy. The aim of this research is to illustrate how object-based method can be applied to the available data to accurately find out vegetation, which can be further sub-classified to obtain area under tree canopy. The result thus obtained gives area under tree canopy with an accuracy of 92.43 % and a Kappa coefficient of 0.80.
Goal location prediction based on deep learning using RGB-D camerajournalBEEI
In the navigation system, the desired destination position plays an essential role since the path planning algorithms takes a current location and goal location as inputs as well as the map of the surrounding environment. The generated path from path planning algorithm is used to guide a user to his final destination. This paper presents a proposed algorithm based on RGB-D camera to predict the goal coordinates in 2D occupancy grid map for visually impaired people navigation system. In recent years, deep learning methods have been used in many object detection tasks. So, the object detection method based on convolution neural network method is adopted in the proposed algorithm. The measuring distance between the current position of a sensor and the detected object depends on the depth data that is acquired from RGB-D camera. Both of the object detected coordinates and depth data has been integrated to get an accurate goal location in a 2D map. This proposed algorithm has been tested on various real-time scenarios. The experiments results indicate to the effectiveness of the proposed algorithm.
Automatic Detection of Window Regions in Indoor Point Clouds Using R-CNNZihao(Gerald) Zhang
The document describes a method to automatically detect window regions in 3D point cloud data of indoor environments collected using a backpack sensor system. The method is based on R-CNN and uses MCG to generate region proposals, extracts features from proposals using a CNN, and classifies proposals as windows or non-windows using a random forest. Experiments on a dataset of 400 images achieved an F1 score of 89.79% and mAP of 96.64% for window detection, outperforming an existing method. Adding a small amount of manually labeled data further improved results.
An effective RGB color selection for complex 3D object structure in scene gra...IJECEIAES
Our goal of the project is to develop a complete, fully detailed 3D interactive model of the human body and systems in the human body, and allow the user to interacts in 3D with all the elements of that system, to teach students about human anatomy. Some organs, which contain a lot of details about a particular anatomy, need to be accurately and fully described in minute detail, such as the brain, lungs, liver and heart. These organs are need have all the detailed descriptions of the medical information needed to learn how to do surgery on them, and should allow the user to add careful and precise marking to indicate the operative landmarks on the surgery location. Adding so many different items of information is challenging when the area to which the information needs to be attached is very detailed and overlaps with all kinds of other medical information related to the area. Existing methods to tag areas was not allowing us sufficient locations to attach the information to. Our solution combines a variety of tagging methods, which use the marking method by selecting the RGB color area that is drawn in the texture, on the complex 3D object structure. Then, it relies on those RGB color codes to tag IDs and create relational tables that store the related information about the specific areas of the anatomy. With this method of marking, it is possible to use the entire set of color values (R, G, B) to identify a set of anatomic regions, and this also makes it possible to define multiple overlapping regions.
Object extraction using edge, motion and saliency information from videoseSAT Journals
Abstract Object detection is a process of finding the instances of object of a certain class which is useful in analysis of video or image. There are number of algorithms have been developed so far for object detection. Object detection has got significant role in variety of areas of computer vision like video surveillance, image retrieval`. In this paper presented an efficient algorithm for moving object extraction using edge, motion and saliency information from videos. Out methodology includes 4 stages: Frame generation, Pre-processing, Foreground generation and integration of cues. Foreground generation includes edge detection using sobel edge detection algorithm, motion detection using pixel-based absolute difference algorithm and motion saliency detection. Conditional Random Field (CRF) is applied for integration of cues and thus we get better spatial information of segmented object. Keywords: Object detection, Saliency information, Sobel edge detection, CRF.
This document presents a methodology for real-time object tracking using a webcam. It combines Prewitt edge detection for object detection and Kalman filtering for tracking. Prewitt edge detection is used to detect the edges of the moving object in each video frame. Then, Kalman filtering is used to track the detected object across subsequent frames by predicting its location. Experiments show the approach can efficiently track objects under deformation, occlusion, and can track multiple objects simultaneously. The combination of Prewitt edge detection and Kalman filtering provides an effective method for real-time object tracking.
Automatic Building detection for satellite Images using IGV and DSMAmit Raikar
This document presents a method for automatic building detection from satellite images using internal gray variance (IGV) and digital surface model (DSM). The proposed method aims to detect low-rising buildings and buildings with partially bright and partially dark rooftops more accurately than existing methods. The key steps include image enhancement, IGV feature extraction, seed point detection using the enhanced image and IGV, clustering using DSM data, binarization, thinning, shadow detection, and segmentation. Results on test satellite images show the method achieves higher detection percentages and lower branch factors than an existing method.
The document summarizes a research paper on using CNN and LSTM models for land use and land cover change detection from remote sensing images. It presents the methodology which involves preprocessing two QuickBird satellite images from Russia from 2005 and 2010. Features are extracted from the segmented images using Gabor, Zernike moments and local features. A CNN model is used for classification and identifies changes such as increased buildings and conversion from forest to land. Performance is evaluated using accuracy, sensitivity, specificity and precision metrics which are over 95% according to the results. The paper concludes the CNN model effectively identifies temporal changes in land use from the satellite images.
This document describes a new method for 2D and 3D object detection and classification for autonomous vehicles using LiDAR and camera (CCD) sensors. It proposes generating object proposals from LiDAR point cloud data by filtering points, projecting them to 2D, and segmenting edges. The proposals are then classified using R-FCN neural network on the CCD image. Class labels from R-FCN are mapped back to edge points to determine the 3D orientation and expand the bounding box for occluded regions. Evaluation on KITTI dataset shows it achieves accurate and fast object detection compared to previous methods.
This document describes a new method for object detection in autonomous vehicles using both LiDAR and image data. The method generates 2D object proposals from LiDAR point cloud data by filtering and grouping edge points. These proposals are then classified using an R-FCN neural network. The class labels are mapped back to the 3D LiDAR points to determine the full 3D bounding box orientations of detected objects. This allows for spatial information and handling of occluded regions, improving over methods using image data alone. The method is evaluated on the KITTI dataset and shown to achieve accurate and fast object detection suitable for autonomous driving applications.
This document provides an overview of image processing techniques for traffic applications. It discusses automatic lane finding using color-based and texture-based segmentation as well as feature-driven approaches. Object detection methods like thresholding, edge detection using Canny operator, and background differencing are also covered. Additionally, the document proposes an emergency vehicle detection system using red beacon detection and frequency analysis to override normal traffic light patterns when an emergency vehicle is detected.
Building Extraction from Satellite ImagesIOSR Journals
This document summarizes a method for detecting buildings from satellite images. It begins with edge detection using Canny edge detection to extract lines from the image. These lines are then input into the Hough Transform to generate line segments. Hypothesis generation searches for rectangular structures using these line segments. Hypothesis verification is performed using color segmentation of the HSV image space and shadow detection. The final output is the extracted buildings from the urban areas of the satellite image.
1. Vision-Based Turf classification for an Autonomous mower
MEng in Robotics and Autonomous Vehicles, Capstone Project Report
Prasanna Kumar Sivakumar
School of Integrative Systems and Design
University of Michigan, Ann Arbor
spkumar@umich.edu
ABSTRACT
This report presents a vision–based turf classification
method for an autonomous lawn mower. The task of
segmenting the scene into turf and non–turf regions is
divided into two phases, data extraction phase and turf
classification phase. During the data extraction phase,
the scene in front of the mower is divided into a grid of
uniformly spaced blocks and color and texture informa-
tion is extracted from the blocks. The color of a block is
represented by average hue and texture by Histogram
of Oriented Gradients (HOG) descriptor. In the turf
classification phase, scene is segmented by classifying
each block as turf or non-turf. Turf classification using
k–means clustering and Support Vector Machines (SVM)
is presented with a detailed study of various parameter
choices on classifier performance.
I. INTRODUCTION
This report presents a prototype lawn mowing robot
capable of detecting non–turf regions and obstacles in
the scene using an off–the–shelf vision sensor. The
objective for the lawn mowing robot is to maintain
turf health and aesthetics of the lawn autonomously
by performing complete coverage of a desired area. To
safely achieve this goal, the system must determine and
maintain containment inside of the permitted area and
also avoid undesired regions and obstacles that may
appear in it’s path. While a wire carrying RF signal
can be used to mark the boundaries of the lawn [1], it
would be difficult to manually demarcate all undesired
regions inside the boundary and impossible to know the
obstacles before hand.
There have been several research prototypes as well
as manufactured products developed for robotic lawn
mowing [1] - [4]. The manufactured products available
in the market make use of a contact sensor to detect a
non–turf object1
after collision or run over it if the height
of object is less than the position of the sensor. None
1
from this point onwards the term ‘non–turf object’ will be used
to refer to both non–turf regions and obstacles
Figure 1: Autonomous lawn mower with monocular
camera used in this work
of the platforms are equipped with sensors to detect the
objects before hand and avoid them. Vision sensors are
an attractive solution to this problem for their potential to
eliminate time consuming installation of boundary wires
and perform sensing functions beyond obstacle detection
such as determining and diagnosing turf problems. A
method for classifying a scene into turf and non–turf
regions using an off–the–shelf vision sensor is presented
in this report. In this work, a robotic mower from John
Deere which is shown in Figure 1 is used.
The task of classifying a scene into turf and non-turf
regions is divided into two phases, data extraction and
turf classification. In the data extraction phase, the color
and texture information in the image captured by the
vision sensor are extracted by treating the image as a
dense grid of uniformly spaced blocks. For representing
color the Average Hue of the block in used and for
representing texture, Histogram of Oriented Gradients
(HOG) descriptor is used. In the classification phase,
the extracted data is passed to a classifier to classify
each block as turf or non–turf. Classification using un-
supervised k–means clustering and supervised Support
Vector Machines (SVM) are presented. The performance
of the two classifiers are compared and it is shown that
an SVM classifier with a Radial Basis Function (RBF)
2. Capstone Project Report Prasanna Kumar Sivakumar
kernel produces better performance under varied lighting
conditions. A detailed study of various implementation
choices on classifier performance is also presented.
A. Related work
Autonomous lawn mowing robots have been receiving
attention from the mobile robotics community in the
recent years. Containment inside the boundary and accu-
rate positioning on the lawn have been the most explored
aspects of robotic mowing, so far. Yang et al. [6]
present a vision based localization method to estimate
the position and ensure boundary containment using an
omni–directional camera. Smith et al. [4] present a novel
positioning technique using Ultrasonic beacons. Batavia
et al. [5] present an obstacle detection technique for a
robotic mower using color and 2–D range data. While
their system performs fairly well, even detecting objects
objects at a curvature of 15cm, it uses of the expensive
laser range finder.
Segmentation of a scene into turf and non–turf regions
for a robotic mower is the central theme of this re-
port. So, only scene segmentation methods that have
application in real-time mobile robotics are mentioned
in this section. There have been extensive literature on
color based scene segmentation. See [7] for survey.
Bruce et al. [8] present a novel threshold based fast im-
age segmentation technique using connected component
analysis for a RoboCup application [9]. Their algorithm
operates at 30Hz with fairly high accuracy for a well
structured indoor scene under constant lighting. How-
ever they do not discuss the performance on unstructured
outdoor conditions. Browning et al. [10] discuss a real-
time color based illumination invariant scene segmen-
tation technique for an outdoor soccer robot but they
assume that the scene is made up distinctly colored
objects.
There have also been extensive literature in edge based
object detection techniques in the past decade. See [11]
for a survey. The method presented in this report inherits
features from the sliding–window object detector work
of Dalal et al. [12] which presented the use of Histogram
of Oriented Gradients descriptor. Tu et al. [13] describe
an approach for identifying regions in the scene. Their
approach has been shown to be effective only on text and
faces. Sudderth et al. [14] relate scenes, objects and parts
in a single hierarchical framework, but do not provide
an exact segmentation of the image. Gould et al. [15]
present a hierarchical region–based approach to image
segmentation. However, their method cannot segment
the same object under different backgrounds and often
leaves them segmented into multiple dissimilar pieces.
B. Organization
The rest of the report is organized as follows. Section
II provides an overview of the data extraction process.
Section III provides an overview of turf classification.
Section IV analyses the performance of each of the
classifier and effects of different parameters in real–
time classification. Section V provides implementation
details. The report is finally concluded in section VI with
plans for future work in section VII.
II. OVERVIEW OF DATA EXTRACTION
This section gives an overview of the data extraction
phase which is summarized in figure 2. This method
is based on evaluating local histograms of image gra-
dient orientations and hue level in a dense grid. The
basic idea is that two attributes differentiate turf regions
from a non-turf object, color and texture. Grass is pre-
dominantly green in color and a patch of grass has dense
edge information which most non-grass objects do not.
This is implemented by dividing the image into small
NxN spatial regions (“blocks”), for each block accumu-
lating a 9–D HOG descriptor which is a histogram of
gradient orientations, and a 1–D average hue number.
The combined 10–D entry forms the representation of
the block.
The following two subsections describe the steps in-
volved data extraction process in detail.
A. Image pre-processing
The input image passed by the camera is a 640x480
RGB image. In order to increase the classification speed
(images classified per second) without losing texture
information, the image is downsampled to a factor of the
original size. A Gaussian blur is applied before down-
sampling to remove high frequency edges and avoid
Moire pattern from forming in the downsampled image.
The lawn mowing robot is typically used outdoors, often
times under bright (mid–day) or dark (late–evening)
lighting conditions, which decreases the global contrast
of the image captured. In order to compensate for this
effect, contrast normalization scheme is applied to the
image which effectively spreads out the most frequent
intensity value. Contrast normalization is carried out
after Gaussian smoothing and before downsampling. The
resultant image is downsampled to a smaller size.
B. Data Collection
The feature information is extracted by sliding an
NxN window across the image. Each window location
is transformed to greyscale and HSV color spaces.
From the grayscale block, sum of edge magnitudes at
2
3. Capstone Project Report Prasanna Kumar Sivakumar
Figure 2: Overview of turf classification chain. The input image is first downsampled to half of its original size
and tiled with a grid of uniformly spaced blocks in which average Hue and HOG vectors are extracted. The
concatenated vectors are fed to a classifier (kmeans/SVM) for classification into turf and non–turf.
9 different oerintations (HOG) is calculated and from
the HSV block, average Hue across the window is
calculated and the two are concatenated to get a 10–D
representation of a block.
III. OVERVIEW OF CLASSIFICATION
A. Dataset
Training dataset consists of images of different types
of turf under different lighting conditions and images of
most common non–turf regions and obstacles as shown
in figure 3. The training image dataset is collected such
that a training image representative of grass contains
only grass and no trace of non–grass and similarly
for non–grass. These training images were collected by
driving the robot manually on different turfs under dif-
ferent lighting conditions and manually cropping regions
that are expected to be best representatives of the two
classes. It is important to note that, only the bottom–
half of an image frame was considered for training.
There are two reasons behind this choice. One, due
to saturation, especially under bright lighting condition,
both the color and the texture information are lost in the
part of the image near horizon and samples collected
from those regions might be misleading and two, since
the autonomous lawn mower is a slow mowing robot, it
takes a longer time to reach the area that are near horizon
in the current image frame. A total of 182 images of
different sizes were collected and for a blocksize of
8x8 148,820 labeled samples were extracted without any
overlap.
B. Dimensionality reduction and scaling
Principal Component Analysis (PCA) is performed
on the training data to identify the N most useful edge
orientations. It is important to note that PCA is applied
on the 9–D HOG data and not the 10–D concatenated
data. The reason behind this choice is using 9–D vector
to represent texture and single number to represent color
skews the data away from color and might increase the
false positive rate. For example, dead grass patch which
has the texture of live grass but not the green color
might be classified as grass. In order to prevent this
from happening, the number of dimensions of texture
data is reduced from 10 to 3. Section IV presents
a detailed analysis of effect of number of Principal
Components on False Positive Rate and Classification
Rate (images/second). The resultant data is normalized
to zero mean and unit variance.
C. k–means clustering
k–means clustering is an obvious unsupervised learn-
ing candidate to solve this problem. There is no training
involved in k–means clustering. Just the result of PCA
is used for dimensionality reduction. k-means clustering
(k=2 in this case) tries to group the samples into two
classes iteratively. The problem with this is, even when
the entire scene is turf, k–means clustering tries to
classify the samples (blocks) obtained from the image
into two classes, thus increasing the false negative rate.
In order to solve this problem, calibration and compar-
ison steps are introduced before and after clustering,
respectively.
Calibration is performed at the beginning of every mow
cycle. During the calibration, the lawn mower is posi-
tioned such that the entire scene in front of it is turf and
one frame of image is grabbed. Samples (dimensionality
reduced) obtained from this image are clustered into two
groups and the corresponding cluster centers are saved.
Since both the clusters contain only turf samples, the
mean of the two centers is assumed to be a representative
of turf cluster center for the current mow cycle. The
resulting value is used as initial turf cluster center for the
first frame of the current mow cycle and from then on,
3
4. Capstone Project Report Prasanna Kumar Sivakumar
(a) (b) (c) (d) (e) (f) (g) (h)
Figure 3: Sample images from the training dataset. (a to d) Sample images of turf under different lighting conditions
(e to h) sample images of most common non-grass objects. During data extraction for training, an NxN window
sliding through each of the images from a to d will be labeled ‘1’ at every location and ‘-1’ when sliding through
each of the images from e to h.
the cluster centers of a frame are used as initial estimates
for the subsequent frame. This reduces the number of
iterations considerably.
Comparison is carried after clustering a frame. If both
the cluster centers are close to the grass cluster center of
the previous frame then the clusters are dissolved and
all the samples are labeled ‘1’ and if both the cluster
centers are far away from the grass cluster center of
the previous frame, then all the samples are labeled ‘-
1’. If only one of the centers is close, the clusters are
preserved. Pseudocode of the comparison step is given
in algorithm 1.
Algorithm 1 Pseudocode for comparison step in k–
means clustering
Input: Clusters and centers of current frame
Cg, Cng, Clustg, Clustng, Turf center of previous
frame Cgprev, threshold
1: if l2norm(Cg − Cgprev) < & l2norm(Cng −
Cgprev) < then
2: Clustg = Clustg ∪ Clustng
3: Clustng = ø
4: else if l2norm(Cg − Cgprev) > & l2norm(Cng −
Cgprev) > then
5: Clustng = Clustng ∪ Clustg
6: Clustg = ø
7: end if
IV. OVERVIEW OF RESULTS
A. Datasets and Evaluation methodology
Testing dataset: The classifiers were tested on six
different datasets. Each dataset was collected by driving
the robot under different lighting condition and consists
of 50 fully labeled images. Most of the labeled images
contain both turf and non-turf regions in them. Figure
4 shows a sample image from each dataset. Five out
of the six datasets were collected on outdoor turf under
varied natural lighting conditions and one was collected
on artificial turf under incandescent lights. The images
were manually labeled using a interactive labeling tool
in MATLAB.
Evaluation Methodology: Performance of a classifier
is quantified by plotting Receiver Operating Character-
istic (ROC) curves [16]. ROC curves plot False Positive
Rates, FPR (1− TrueNeg
FalsePos+TrueNeg ) against True Positive
Rates, TPR ( TruePos
TruePos+FalseNeg ). Lower values of FPR
are better. Another evaluation metric is the Classification
rate, CR which is number of 320x240 resolution images
classified per second by the classifier (images/second
or Hz). Higher classification rates are better but since,
a robotic lawn mower is a slow moving robot2
, classifi-
cation rate of 5Hz is considered good.
B. k–means clustering vs Support Vector Machines
Classifier FPR CR (Hz)
k–means 0.2594 26.6746
SVM
(RBF kernel, σ = 1)
0.1036 8.1748
Table I: False positive rates and classification rates for
the two classifiers
k–means clustering produces 15% more false positives
than SVM but classifies up to 3 times more frames per
second. Since, SVM with RBF kernel classifies at CR
of more than expected 5Hz and has better precision
than k–means clustering, SVM classifier is chosen to be
implemented on the mower. Rest of the section presents
the effects of different parameter choices for the SVM
classifier.
C. Effects of n Principal components (nPC)
PCA is applied to 9–D HOG vector to identify n most
useful HOG orientations and the resulting transformed
n–D vector is concatenated with 1–D color vector and
used for training. Bar graph in figure 5 shows FPR
2
maximum ground speed of the robot - 13 cm/sec
4
5. Capstone Project Report Prasanna Kumar Sivakumar
(a) (b) (c) (d) (e) (f)
Figure 4: Sample image from each of the six datasets. a to e were collected outdoor on natural turf and f was
collected on astroturf, indoor, under artificial lighting
Figure 5: FPR and classification time (1/CR) for num-
ber of Principal Components (nPC). FPR decreases from
nPC=1 to 3 and increases afterwards
and time taken to classify a frame of image (1/CR)
for increasing number of principal components. Classi-
fication time increases with the number of components.
The FPR decreases from nPC = 1 to 3 and increases
afterwards. This trend with FPR shows the need for a
proper balance between color and texture information to
handle false positives. When nPC is small (nPC < 3)
the training data is skewed towards color and any non–
turf object which has the green color of lawn but not the
texture (eg: green lawn spread) might be classified as
turf, thus increasing FPR. Similarly, when nPC is large
(nPC > 5) the training data is skewed towards texture
and objects like dead grass patch which has the texture
of live grass might be classified as turf. The right balance
between color and texture is reached at nPC = 3 which
has the lowest FPR of 0.1036.
D. Effects kernel choices
Top plot of figure 6 shows the performance of SVM
classifier for various kernel types. The ROC curves were
plotted for number of Principal Components, nPC = 3.
RBF kernel, defined in equation 1, has the lowest FPR
(0.1036) and it performs significantly better than other
Figure 6: (top) ROC curves for various kernel types and
(bottom) ROC curves for various values of σ for RBF
kernel
kernel types.
K(x, x ) = exp(−
x − x 2
2σ2
) (1)
Bottom plot of figure 6 shows the performance of RBF
kernel SVM for various values of the free parameter, σ.
Although not significant, FPR decreases from σ = 0.25
to σ = 1 and increases after that. σ = 1 has the lowest
FPR of 0.1036.
E. Effects of blocksizes
SVM classifier with RBF kernel (σ = 1 and nPC =
3) was trained and tested for various blocksizes without
5
6. Capstone Project Report Prasanna Kumar Sivakumar
Figure 7: ROC curves for various block sizes used in
data–extraction
overlap. Plot in figure 7 shows the performance of SVM
classifier for different blocksizes. Blocksize of 8x8 has
the lowest FPR of 0.1036.
F. Effect of feature standardization
During training and testing, the feature data extracted
from the images are standardized to zero mean and unit
variance as defined in equation 2. xij is the jth feature
of ith sample and ¯xj and σxj
are the mean and SD for
feature j. Transforming the data to this scale improves
the classifier performance significantly.
xij =
xij − ¯xj
σxj
(2)
For nPC = 3 an unstandardized sample looks like,
< 0.2385, 103.3826, 987.4378, 678.0125 > and SVM
classifier with RBF kernel (σ = 1) has FPR = 0.3216.
After standardizing using equation 2, the same data
looks like, < 0.1543, 0.0178, 0.7621, 0.1675 > and the
same classifier has FPR = 0.1036.
V. IMPLEMENTATION
The vision hardware used is a video camera from
logitech [21]. It has a 1.3 Megapixel CMOS sensor
and captures VGA standard images at a resolution
of 640x480. The camera is mounted on the front of
the robot using a plastic mount (3D printed) which
allows lateral motion and pan tilt. The computer used
in training and implementation runs on a 2Ghz Intel
Pentium Processor with 4GB of RAM. OpenCV [22]
was used for the purpose of data extraction and libSVM
[23] was used in training the SVM classifier.
VI. CONCLUSION
A vision–based turf classification method for an au-
tonomous lawn mower was presented in this report. A
two–phase scene segmentation technique, using color
and texture information, was proposed. During the first,
data–extraction phase, color and texture information
from the image captured by the vision sensor is extracted
by treating the image as a dense grid of uniformly spaced
blocks. During the second, turf classification phase, data
extracted from each block is classified as turf or non–
turf. Performance of two different classifiers, k–means
clustering and Support Vector Machines was presented
with a detailed parameter study for SVM classifier.
A notable finding of the study is the need for proper bal-
ance between the number of color and texture features.
By performing Principal Component Analysis (PCA) on
texture data extracted from training dataset, it was found
that using excessive or lesser number of texture features
results in higher False Positive Rates. It was found that
using 3 features for texture provides the right balance
between texture and color data. Parameter study on the
SVM classifier showed that a RBF kernel with σ = 1
results in the best performance.
VII. FUTURE WORK
Although the current SVM classifier is reasonably
accurate and efficient – processing 320x240 resolution
images at 8Hz, there is still room for improvement and
optimization. One of the glaring omissions of this work
is the removal of Infrared (IR) filter from the camera.
Since chlorophyll in live grass reflects radiations in near
IR region of spectrum [17], removing IR filter would
greatly improve detection of live grass. Using Local
Binary Pattern [18] in combination with HOG descriptor
for feature representation can improve the classification
accuracy. LBP has found to be a powerful feature for
texture classification [19]. Using auto–encoders with
Convolutional Neural Networks [20] can improve the
accuracy and classification time greatly.
ACKNOWLEDGMENTS
This work was funded by Multi-Disciplinary design
Program (MDP) of University of Michigan and sup-
ported by John Deere. In particular, I thank and ac-
knowledge Mr. David Johnson of John Deere Advanced
R&D for his support and valuable inputs, Prof. Matthew
Johnson Roberson for technical guidance throughout the
project and Mr. Daniel Kline for guiding the completion
of the project from a product development standpoint.
6
7. Capstone Project Report Prasanna Kumar Sivakumar
REFERENCES
[1] ConsumerSearch, “Best robotic mowers”, [Online]. Available:
http://www.consumersearch.com/robotic-lawnmowers/best-
robotic-mowers
[2] R. W. Hicks II and E. L. Hall, “Survey of robot lawn mowers”,
in Proc. SPIE Intelligent Robots and Computer Vision XIX: Al-
gorithms, Techniques, and Active Vision, Oct. 2000, pp. 262269.
[Online]. Available: http://dx.doi.org/10.1117/12.403770
[3] H. Sahin and L. Guvenc, “Household robotics: autonomous de-
vices for vacuuming and lawn mowing [applications of control]”,
IEEE Control Systems, vol. 27, no. 2, pp. 2096, 2007.
[4] A. Smith, H. Chang, and E. Blanchard, “An outdoor high-
accuracy local positioning system for an autonomous robotic
golf greens mower”, in Proc. IEEE International Conference on
Robotics and Automation, May 2012, pp. 26332639.
[5] P. Batavia and S. Singh, “Obstacle Detection in Smooth High-
Curvature Terrain”, Proc. IEEE International Conference on
Robotics and Automation, May 2002, pp. 1283-1291.
[6] J. Yang, s. Chung, S. Hutchinson, D. Johnson, and M. Kise,
“Vision-Based Localization and Mapping for an Autonomous
Mower”, in Proc. IEEE International Conference on Intelligent
Robots and Systems, November 2013, pp. 20182026.
[7] W. Skarbek ans A. Koschan, “Colour Image Segmentation, A
Survey”, Online Library Technical University of Berlin, Oct.
1994. [Online] Available: http://iristown.engr.utk.edu/ koschan/-
paper/coseg.pdf.
[8] J. Bruce, T. Balch and M. Veloso, “Fast and Inexpensive Color
Image Segmentation for Interactive Robots”, in Proc. IEEE
International Conference on Intelligent Robots and Systems,
November 2000, pp. 10181026.
[9] H. Kitano, Y. Kuniyoshi, I. Noda, M. Asada, H. Matsubara, and
E. Osawa. RoboCup: A challenge problem for AI. AI Magazine,
18(1), pages 7385, 1997.
[10] B. Browning and D. Govindaraju, “Fast, Robust Techniques
for colored object Detection in Variable Lighting Conditions”,
in Proc. IEEE International Conference on Robotics and Au-
tomation, May 2004, pp. 16631669.
[11] D. M. Gavrila. “The visual analysis of human movement: A sur-
vey. Computer Vision and Image Understanding”, International
Journal on Computer Vision, vol. 73, pp. 8298, 1999.
[12] N. Dalal and B. Triggs, “Histograms of Oriented Gradients for
Human Detection”, In Proc. Computer Society Conference on
Computer Vision and Pattern Recognition, February 2005, pp.
1063-1071.
[13] Z. Tu, X. Chen, A. L. Yuille and S.-C. Zhu, “Image parsing:
Unifying segmentation’, detection, and recognition”. In Proc.
International Conference on Computer Vision, May 2003, pp.
928-934.
[14] E. Sudderth, A. Torralba, W. Freeman, and A. Willsky, “De-
scribing visual scenes using transformed objects and parts”,
International Journal on Computer Vision, Vol. 87, pp. 423-431,
2007.
[15] S. Gould, T. Gao and D. Koller, “Region-based Segmentation
and Object Detection”, In Proc. International Conference on
Neural Information Processing Systems, December 2009, pp.
1111-1120.
[16] T. Fawcett, “An Introduction to ROC Analysis”. Pattern Recog-
nition Letters 27 (8): 861874. doi:10.1016/j.patrec.2005.10.010
[17] Science Mission Directorate, “Reflected Near-
Infrared Waves”, [Online] Mission:Science, Na-
tional Aeronautics and Space Administration.
http://missionscience.nasa.gov/ems/08 nearinfraredwaves.html.
[18] DC. He and L. Wang (1990), ”Texture Unit, Texture Spectrum,
And Texture Analysis”, Geoscience and Remote Sensing, IEEE
Transactions on, vol. 28, pp. 509 - 512.
[19] X. Wang, T. Han and S. Yan, “An HOG-LBP Human Detector
with Partial Occlusion Handling”, In Proc. International Confer-
ence on Computer Vision, May 2009, pp. 456-464.
[20] Bengio, Y, “Learning Deep Architectures for AI”, Foundations
and Trends in Machine Learning, vol.06, December 2009.
[21] Logitech Webcam C110, [online]
http://support.logitech.com/product/8112.
[22] G. Bradski, “OpenCV, Open Source Computer Vision Library”,
Dr. Dobb’s Journal of Software Tools, Volume 4, 2008.
[23] C. Chang and C. Lin, “LIBSVM: Library for support vec-
tor machines”, ACM Transactions on Intelligent Systems and
Technology, Volume 2, Issue 3, pp. 27:1–27:27, 2011. Software
available at http://www.csie.ntu.edu.tw/ cjlin/libsvm.
7