SlideShare a Scribd company logo
1 of 44
GARBY BABY
MCA1414
NIRMALA COLLEGE
MUVATTUPUZHA
1
INTRODUCTION
 Today the most information is available either
on paper or in the form of photographs or
videos.
 Large information is stored in images.
 The current technology is restricted to extracting
text against clean backgrounds.
 Thus, there is a need for a system to extract text
from general backgrounds.
2
 Text Extraction and recognition in Images has
become a potential application in many fields
like Image indexing , Robotics, Intelligent
transport systems etc.
 For example capturing license plate information
through a video camera and extracting license
number in traffic signals
 However, variations of text due to differences in
size, style, orientation, and alignment, as well as
low image contrast and complex background
make the problem of automatic text extraction
extremely challenging.
3
 Content-based image indexing refers to the
process of attaching labels to images based on
their content
 Image content can be divided into two main
categories:
1) Perceptual content and
2)semantic content
Perceptual content includes attributes such as
Color , intensity, shape, texture, and their
temporal changes,
4
 A number of studies on the use of relatively low-
level perceptual content for image have already
been reported
 Semantic content means objects, events, and
their relations.
 Studies on semantic image content in the form
of text, face, vehicle, and human action have
also attracted some recent interest.
5
 Among them, text within an image is of particular
interest as
(i) It is very useful for describing the contents of
an image.
(ii)It can be easily extracted compared to other
semantic contents, and
(iii)It enables applications such as
keyword-based image search, and
text-based image indexing.
6
7
TEXT IN IMAGES
 A variety of approaches to Text Information
Extraction (TIE) from images have been
proposed for specific applications including page
segmentation , address block location , license
plate location , and content-based image/video
indexing
 In spite of such extensive studies, it is still not
easy to design a general-purpose TIE system.
8
 This is because there are so many possible
sources of variation when extracting text from a
shaded or textured background, from low-
contrast or complex images, or from images
having variations in font size, style, color ,
orientation, and alignment.
 These variations make the problem of automatic
TIE extremely difficult.
 Figures 1-4 show some examples of text in
images.
9
 (a) (b) ©
 Fig. 1. Grays cale document images: (a) single-
column text from a book,
 (b) two-column page from a journal (IEEE
Transactions on PAMI), and
 (c) an electrical drawing (courtesy of Lu [24]).
10
(Fig. 2), Multi-colour document images .
 (a) (b) ©
 Fig. 2. Multi-color document images: each text
line may or may not be of the same color.
11
Text in video images can be further classified into
caption text (Fig. 3), or scene text (Fig. 4).
 Fig.3 (Images with caption text)
Text is artificially overlaid on the image.
(a)
(a) Shows captions overlaid directly on the
background
12
 Fig. 4. Scene text images: Images with
variations in skew, perspective, blur, illumination,
and alignment.
13
TERMS
 Before we attempt to explain the various
techniques used in TIE, it is important to define
the commonly used terms and summarize the
characteristics of text that can be used for TIE
algorithms.
 Text in images can exhibit many variations with
respect to the following properties.
14
1. Geometry
 Size: Although the text size can vary a lot,
assumptions can be made depending on the
application domain.
 Alignment: The characters in the caption text
appear in clusters and usually lie horizontally.
 Although sometimes they can appear as
non- planar texts as a result of special effects.
 This does not apply to scene text, which can have
various perspective distortions. Scene text can be
aligned in any direction and can have geometric
distortions (Fig. 4).
15
 Inter-character distance: characters in a text
line have a uniform distance between them.
2. Colour
 The characters in a text line tend to have the
same or similar colours.
 This property makes it possible to use a
connected component-based approach for text
detection.
 Most of the research reported till date has
concentrated on finding ‘text strings of a single
colour (monochrome)’.
16
 However, video images and other complex
colour documents can contain ‘text strings with
more than two colours (polychrome)’ for effective
visualization, i.e., different colours within one
word.
3. Motion
 The same characters usually exist in
consecutive frames in a video with or without
movement.
 This property is used in text tracking and
enhancement.
 Caption text usually moves in a uniform way:
horizontally or vertically
17
 Scene text can have arbitrary motion due to
camera or object movement.
4. Edge
 Most caption and scene text are designed to be
easily read, thereby resulting in strong edges at
the boundaries of text and background
5. Compression
 Many digital images are recorded, transferred,
and processed in a compressed format.
 Thus, a faster TIE system can be achieved if
one can extract text without decompression
18
 The role of text detection is to find the image
regions containing only text that can be directly
highlighted to the user or fed into an optical
character reader module for recognition.
 In this seminar a new system is discussing
which extracts text in images.
 The system takes colored images as input.
 It detects text on the basis of certain text
features:
 The image is then cleaned up so that the text
stands out.
19
RELATED WORK
 Various methods have been proposed in the
past for detection and localization of text in
images and videos.
 These approaches take into consideration
different properties related to text in an image
such as color, intensity, connected-components,
edges etc.
20
 These properties are used to distinguish text
regions from their background and/or other
regions within the image.
 In the algorithm based on color clustering, the
input image is first pre-processed to remove any
noise if present.
 Then the image is grouped into different color
layers and a gray component.
 This approach utilizes the fact that usually the
color data in text characters is different from the
color data in the background.
21
 The potential text regions are localized using
connected component based heuristics from
these layers.
 The experiments conducted show that the
algorithm is robust in locating mostly Chinese
and English characters in images.
 The text detection algorithm is also based on
color continuity.
 In addition it also uses multi-resolution wavelet
transforms and combines low as well as high
level image features for text region extraction.
22
 Texture based segmentation is used to
distinguish text from its background.
 Further a bottom-up „chip generation‟ process
is carried out which uses the spatial cohesion
property of text characters.
 The chips are collections of pixels in the image
consisting of potential text strokes and edges.
 The results show that the algorithm is robust in
most cases, except for very small text characters
that are not properly detected.
 Also in the case of low contrast in the image,
misclassifications occur in the texture
segmentation.
23
METHOD
Various steps involved in TIE process are
(1) Converting colored image to grayscale
(2) Binarization
(3) Connected components
(4) Horizontal and Vertical Projections
(5) Reconstruction.
(6) Recognition using OCR
24
 Converting colored image to grayscale
 A digital color image is a color image that
includes color information for each pixel.
 There are various color models which are used
to represent a color image.
 These are RGB color model, in which red, green
and blue light is added together in various ways
to reproduce a broad array of colors.
25
 The other models are CMY color model which
uses cyan, magenta and yellow light and HSI
model which uses hue, saturation and intensity
variations.
 Grayscale images have range of shades of gray
without apparent color.
 These are used as less information needs to be
provided for each pixel.
 In an 8 bit image, 256 shades are possible.
 The darkest possible shade black is
represented as 00000000 and lightest possible
shade white is represented as 11111111.
26
 Binarization
 A Binary image is a digital image that can have only
two possible values for each pixel.
 Each pixel is stored as single bit 0 or 1.
 The name black and white is often used for this
concept.
 To form a binary image we select a threshold
intensity value.
27
 All the pixels having intensity greater than the
threshold value are changes to 0 (black) and the
pixels with intensity value less than the threshold
intensity value are changed to 1 (white).
 Thus the image is changed to a binary image .
Connected components
 For two pixels to be connected they must be
neighbors and their gray levels must specify a
certain criterion of similarity.
28
 For example, in a binary image with values 0
and 1, two pixels may be neighbors but they are
said to be connected only if they have same
values.
 A pixel p with coordinates (x, y) has four
horizontal and vertical neighbors known as 4-
neighbors of p, given as: (x, y+1), (x, y-1), (x+1,
y), (x-1, y) and four diagonal neighbors given as:
(x+1, y-1), (x-1, y-1), (x+1, y+1), (x-1, y+1).
 Together these are known as 8-neighbors of p.
29
 If S represents subset of pixels in an image, two
pixels p and q are said to be connected if there
exists a path between them consisting entirely of
pixels in S.
 For any pixel p in S, the set of pixels that are
connected to it in S is called a connected
component of S.
30
Horizontal and Vertical Projections
 The method is performed on binary images.
 It starts scanning from left side of every line and
records a change in case of facing the pixel
change from zero to one and again to zero.
 Counting the change does not depend on
number of pixels in this method.
 Robustness in noisy condition is an advantage
of this method.
 After finding the connected components, we
check the transitions in the values of pixels
horizontally.
31
 Transitions can be either zero (black) to one (white)
or one (white) to zero (black).
 The text regions will have larger number of
transitions from black to white or vice versa,
whereas the background region will have lesser
number of transitions.
 If the allocated amount of changes for each row is
between two thresholds (low and high thresholds),
the row potentially would be considered as text area
and the up and down of this row would be specified.
32
 Next, we search vertically for finding the exact
location of the text and ignoring these rows as a
text.
 For finding the exact location of the text, we use
some heuristics.
 These heuristics include height and length of the
text and the ratio of height to length and enough
number of pixels in this area.
33
Reconstruction
 After the extraction of text regions from images,
the text regions become a bit distorted and
difficult to read, thus we recover these
components using the original image.
 The distorted and original images are compared
with each other and the pixels which are erased
or disfigured are recovered.
34
Text Extraction Results
35
Recognition using OCR
 The OCR software converts each character into
ASCII codes.
 The extracted text is stored as ASCII codes in
the computer memory.
 OCR systems have been available for a number
of years and the current commercial systems
can produce an extremely high recognition rate
for machine-printed documents on a simple
background.
36
 Applications
 There are numerous applications of a text
information extraction system, including
document analysis, vehicle license plate
extraction, technical paper analysis, and object-
oriented data compression.
 In the following, we briefly describe some of
these applications
37
 .
 Wearable or portable computers: with the
rapid development of computer hardware
technology, wearable computers are now a
reality.
 A TIE system involving a hand-held device and
camera was presented as an application of a
wearable vision system.
 Translation camera can detect text in a scene
image and translate Japanese text into English
after performing character recognition.
38
 Content-based video coding or document
coding: The MPEG-4 standard supports object-
based encoding.
 When text regions are segmented from other
regions in an image, this can provide higher
compression rates and better image quality.
 As a result, they can achieve a higher quality
rendering of documents containing text, pictures,
and graphics.
 Text-based image indexing: This involves
automatic text-based video structuring methods
using caption data .
39
 License/container plate recognition: There
has already been a lot of work done on vehicle
license plate and container plate recognition.
 Although container and vehicle license plates
share many characteristics with scene text,
many assumptions have been made regarding
the image acquisition process (camera and
vehicle position and direction, illumination,
character types, and color) and geometric
attributes of the text.
40
 Texts in WWW images: The extraction of text
from WWW images can provide relevant
information on the Internet.
 Video content analysis: Extracted text regions
or the output of character recognition can be
useful in genre recognition .
 The size, position, frequency, text alignment,
and OCR-ed results can all be used for this.
 Industrial automation: Part identification can
be accomplished by using the text information
on each part.
41
CONCLUSION
 Current technologies do not work well for
documents with text printed against shaded or
textured backgrounds or those with non-
structured layout.
 In contrast, our technique works well for normal
documents as well as with the documents
described above situations.
 The system is stable and robust. All the system
parameters remain the same throughout all the
experiments.
42
 However, it is not easy to use commercial OCR
software for recognizing text extracted from
images or video frames.
 New OCR systems need to be developed to
handle large amount of noise and distortion in
TIE applications.
43
44

More Related Content

What's hot

Feature Extraction
Feature ExtractionFeature Extraction
Feature Extractionskylian
 
Handwritten character recognition using artificial neural network
Handwritten character recognition using artificial neural networkHandwritten character recognition using artificial neural network
Handwritten character recognition using artificial neural networkHarshana Madusanka Jayamaha
 
Detecting text from natural images with Stroke Width Transform
Detecting text from natural images with Stroke Width TransformDetecting text from natural images with Stroke Width Transform
Detecting text from natural images with Stroke Width TransformPooja G N
 
Image Representation & Descriptors
Image Representation & DescriptorsImage Representation & Descriptors
Image Representation & DescriptorsPundrikPatel
 
Region based segmentation
Region based segmentationRegion based segmentation
Region based segmentationramya marichamy
 
Cat and dog classification
Cat and dog classificationCat and dog classification
Cat and dog classificationomaraldabash
 
Optical Character Recognition (OCR) based Retrieval
Optical Character Recognition (OCR) based RetrievalOptical Character Recognition (OCR) based Retrieval
Optical Character Recognition (OCR) based RetrievalBiniam Asnake
 
PPT on BRAIN TUMOR detection in MRI images based on IMAGE SEGMENTATION
PPT on BRAIN TUMOR detection in MRI images based on  IMAGE SEGMENTATION PPT on BRAIN TUMOR detection in MRI images based on  IMAGE SEGMENTATION
PPT on BRAIN TUMOR detection in MRI images based on IMAGE SEGMENTATION khanam22
 
OCR (Optical Character Recognition)
OCR (Optical Character Recognition) OCR (Optical Character Recognition)
OCR (Optical Character Recognition) IstiaqueBinIslam
 
Detection and recognition of face using neural network
Detection and recognition of face using neural networkDetection and recognition of face using neural network
Detection and recognition of face using neural networkSmriti Tikoo
 
Character Recognition using Machine Learning
Character Recognition using Machine LearningCharacter Recognition using Machine Learning
Character Recognition using Machine LearningRitwikSaurabh1
 
CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...
CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...
CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...Editor IJMTER
 
Attributes of output primitive(line attributes)
Attributes of output primitive(line attributes)Attributes of output primitive(line attributes)
Attributes of output primitive(line attributes)shalinikarunakaran1
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Gaurav Mittal
 
Classification and Regression
Classification and RegressionClassification and Regression
Classification and RegressionMegha Sharma
 

What's hot (20)

Feature Extraction
Feature ExtractionFeature Extraction
Feature Extraction
 
Image segmentation
Image segmentation Image segmentation
Image segmentation
 
Handwritten character recognition using artificial neural network
Handwritten character recognition using artificial neural networkHandwritten character recognition using artificial neural network
Handwritten character recognition using artificial neural network
 
Detecting text from natural images with Stroke Width Transform
Detecting text from natural images with Stroke Width TransformDetecting text from natural images with Stroke Width Transform
Detecting text from natural images with Stroke Width Transform
 
Image Representation & Descriptors
Image Representation & DescriptorsImage Representation & Descriptors
Image Representation & Descriptors
 
Region based segmentation
Region based segmentationRegion based segmentation
Region based segmentation
 
Bresenham algorithm
Bresenham algorithmBresenham algorithm
Bresenham algorithm
 
Region based segmentation
Region based segmentationRegion based segmentation
Region based segmentation
 
Cat and dog classification
Cat and dog classificationCat and dog classification
Cat and dog classification
 
Optical Character Recognition (OCR) based Retrieval
Optical Character Recognition (OCR) based RetrievalOptical Character Recognition (OCR) based Retrieval
Optical Character Recognition (OCR) based Retrieval
 
PPT on BRAIN TUMOR detection in MRI images based on IMAGE SEGMENTATION
PPT on BRAIN TUMOR detection in MRI images based on  IMAGE SEGMENTATION PPT on BRAIN TUMOR detection in MRI images based on  IMAGE SEGMENTATION
PPT on BRAIN TUMOR detection in MRI images based on IMAGE SEGMENTATION
 
OCR (Optical Character Recognition)
OCR (Optical Character Recognition) OCR (Optical Character Recognition)
OCR (Optical Character Recognition)
 
Lzw coding technique for image compression
Lzw coding technique for image compressionLzw coding technique for image compression
Lzw coding technique for image compression
 
Data Redundacy
Data RedundacyData Redundacy
Data Redundacy
 
Detection and recognition of face using neural network
Detection and recognition of face using neural networkDetection and recognition of face using neural network
Detection and recognition of face using neural network
 
Character Recognition using Machine Learning
Character Recognition using Machine LearningCharacter Recognition using Machine Learning
Character Recognition using Machine Learning
 
CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...
CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...
CHARACTER RECOGNITION USING NEURAL NETWORK WITHOUT FEATURE EXTRACTION FOR KAN...
 
Attributes of output primitive(line attributes)
Attributes of output primitive(line attributes)Attributes of output primitive(line attributes)
Attributes of output primitive(line attributes)
 
Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)Convolutional Neural Networks (CNN)
Convolutional Neural Networks (CNN)
 
Classification and Regression
Classification and RegressionClassification and Regression
Classification and Regression
 

Viewers also liked

Text extraction from natural scene image, a survey
Text extraction from natural scene image, a surveyText extraction from natural scene image, a survey
Text extraction from natural scene image, a surveySOYEON KIM
 
optical character recognition system
optical character recognition systemoptical character recognition system
optical character recognition systemVijay Apurva
 
Optical Character Recognition( OCR )
Optical Character Recognition( OCR )Optical Character Recognition( OCR )
Optical Character Recognition( OCR )Karan Panjwani
 
Extracting text form video_using_matlab
Extracting text form video_using_matlabExtracting text form video_using_matlab
Extracting text form video_using_matlabPula Likhitha
 
Hand Written Character Recognition Using Neural Networks
Hand Written Character Recognition Using Neural Networks Hand Written Character Recognition Using Neural Networks
Hand Written Character Recognition Using Neural Networks Chiranjeevi Adi
 

Viewers also liked (6)

Text extraction from natural scene image, a survey
Text extraction from natural scene image, a surveyText extraction from natural scene image, a survey
Text extraction from natural scene image, a survey
 
optical character recognition system
optical character recognition systemoptical character recognition system
optical character recognition system
 
Optical Character Recognition( OCR )
Optical Character Recognition( OCR )Optical Character Recognition( OCR )
Optical Character Recognition( OCR )
 
Textmining Introduction
Textmining IntroductionTextmining Introduction
Textmining Introduction
 
Extracting text form video_using_matlab
Extracting text form video_using_matlabExtracting text form video_using_matlab
Extracting text form video_using_matlab
 
Hand Written Character Recognition Using Neural Networks
Hand Written Character Recognition Using Neural Networks Hand Written Character Recognition Using Neural Networks
Hand Written Character Recognition Using Neural Networks
 

Similar to Text Extraction from Images Using Color and Geometry Features

Block Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid ExtensionBlock Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid ExtensionDR.P.S.JAGADEESH KUMAR
 
DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...
DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...
DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...Mohammad Liton Hossain
 
A comparative study on content based image retrieval methods
A comparative study on content based image retrieval methodsA comparative study on content based image retrieval methods
A comparative study on content based image retrieval methodsIJLT EMAS
 
A comparative analysis of retrieval techniques in content based image retrieval
A comparative analysis of retrieval techniques in content based image retrievalA comparative analysis of retrieval techniques in content based image retrieval
A comparative analysis of retrieval techniques in content based image retrievalcsandit
 
A COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVAL
A COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVALA COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVAL
A COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVALcscpconf
 
Text Extraction of Colour Images using Mathematical Morphology & HAAR Transform
Text Extraction of Colour Images using Mathematical Morphology & HAAR TransformText Extraction of Colour Images using Mathematical Morphology & HAAR Transform
Text Extraction of Colour Images using Mathematical Morphology & HAAR TransformIOSR Journals
 
AVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification SchemeAVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification SchemeDR.P.S.JAGADEESH KUMAR
 
A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosCSCJournals
 
A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosCSCJournals
 
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...IJERA Editor
 
An Unsupervised Cluster-based Image Retrieval Algorithm using Relevance Feedback
An Unsupervised Cluster-based Image Retrieval Algorithm using Relevance FeedbackAn Unsupervised Cluster-based Image Retrieval Algorithm using Relevance Feedback
An Unsupervised Cluster-based Image Retrieval Algorithm using Relevance FeedbackIJMIT JOURNAL
 
Information search using text and image query
Information search using text and image queryInformation search using text and image query
Information search using text and image queryeSAT Journals
 
Information search using text and image query
Information search using text and image queryInformation search using text and image query
Information search using text and image queryeSAT Publishing House
 
Survey on Content Based Image Retrieval
Survey on Content Based Image Retrieval Survey on Content Based Image Retrieval
Survey on Content Based Image Retrieval ijcax
 

Similar to Text Extraction from Images Using Color and Geometry Features (20)

E1803012329
E1803012329E1803012329
E1803012329
 
Block Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid ExtensionBlock Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid Extension
 
DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...
DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...
DEVELOPMENT OF AN ALPHABETIC CHARACTER RECOGNITION SYSTEM USING MATLAB FOR BA...
 
Ac03401600163.
Ac03401600163.Ac03401600163.
Ac03401600163.
 
A comparative study on content based image retrieval methods
A comparative study on content based image retrieval methodsA comparative study on content based image retrieval methods
A comparative study on content based image retrieval methods
 
A comparative analysis of retrieval techniques in content based image retrieval
A comparative analysis of retrieval techniques in content based image retrievalA comparative analysis of retrieval techniques in content based image retrieval
A comparative analysis of retrieval techniques in content based image retrieval
 
A COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVAL
A COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVALA COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVAL
A COMPARATIVE ANALYSIS OF RETRIEVAL TECHNIQUES IN CONTENT BASED IMAGE RETRIEVAL
 
C04741319
C04741319C04741319
C04741319
 
Text Extraction of Colour Images using Mathematical Morphology & HAAR Transform
Text Extraction of Colour Images using Mathematical Morphology & HAAR TransformText Extraction of Colour Images using Mathematical Morphology & HAAR Transform
Text Extraction of Colour Images using Mathematical Morphology & HAAR Transform
 
AVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification SchemeAVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification Scheme
 
A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In Videos
 
A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In Videos
 
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
 
An Unsupervised Cluster-based Image Retrieval Algorithm using Relevance Feedback
An Unsupervised Cluster-based Image Retrieval Algorithm using Relevance FeedbackAn Unsupervised Cluster-based Image Retrieval Algorithm using Relevance Feedback
An Unsupervised Cluster-based Image Retrieval Algorithm using Relevance Feedback
 
Information search using text and image query
Information search using text and image queryInformation search using text and image query
Information search using text and image query
 
Fc4301935938
Fc4301935938Fc4301935938
Fc4301935938
 
Information search using text and image query
Information search using text and image queryInformation search using text and image query
Information search using text and image query
 
Survey on Content Based Image Retrieval
Survey on Content Based Image Retrieval Survey on Content Based Image Retrieval
Survey on Content Based Image Retrieval
 
50320140502001 2
50320140502001 250320140502001 2
50320140502001 2
 
50320140502001
5032014050200150320140502001
50320140502001
 

Recently uploaded

Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Allon Mureinik
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsMemoori
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking MenDelhi Call girls
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxMalak Abu Hammad
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationMichael W. Hawkins
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAndikSusilo4
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Hyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your Budget
Hyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your BudgetHyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your Budget
Hyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your BudgetEnjoy Anytime
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphNeo4j
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):comworks
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsHyundai Motor Group
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 

Recently uploaded (20)

Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)Injustice - Developers Among Us (SciFiDevCon 2024)
Injustice - Developers Among Us (SciFiDevCon 2024)
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial Buildings
 
08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men08448380779 Call Girls In Friends Colony Women Seeking Men
08448380779 Call Girls In Friends Colony Women Seeking Men
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 
The transition to renewables in India.pdf
The transition to renewables in India.pdfThe transition to renewables in India.pdf
The transition to renewables in India.pdf
 
The Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptxThe Codex of Business Writing Software for Real-World Solutions 2.pptx
The Codex of Business Writing Software for Real-World Solutions 2.pptx
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
Azure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & ApplicationAzure Monitor & Application Insight to monitor Infrastructure & Application
Azure Monitor & Application Insight to monitor Infrastructure & Application
 
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptxVulnerability_Management_GRC_by Sohang Sengupta.pptx
Vulnerability_Management_GRC_by Sohang Sengupta.pptx
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Hyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your Budget
Hyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your BudgetHyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your Budget
Hyderabad Call Girls Khairatabad ✨ 7001305949 ✨ Cheap Price Your Budget
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
 
CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):CloudStudio User manual (basic edition):
CloudStudio User manual (basic edition):
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter RoadsSnow Chain-Integrated Tire for a Safe Drive on Winter Roads
Snow Chain-Integrated Tire for a Safe Drive on Winter Roads
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 

Text Extraction from Images Using Color and Geometry Features

  • 2. INTRODUCTION  Today the most information is available either on paper or in the form of photographs or videos.  Large information is stored in images.  The current technology is restricted to extracting text against clean backgrounds.  Thus, there is a need for a system to extract text from general backgrounds. 2
  • 3.  Text Extraction and recognition in Images has become a potential application in many fields like Image indexing , Robotics, Intelligent transport systems etc.  For example capturing license plate information through a video camera and extracting license number in traffic signals  However, variations of text due to differences in size, style, orientation, and alignment, as well as low image contrast and complex background make the problem of automatic text extraction extremely challenging. 3
  • 4.  Content-based image indexing refers to the process of attaching labels to images based on their content  Image content can be divided into two main categories: 1) Perceptual content and 2)semantic content Perceptual content includes attributes such as Color , intensity, shape, texture, and their temporal changes, 4
  • 5.  A number of studies on the use of relatively low- level perceptual content for image have already been reported  Semantic content means objects, events, and their relations.  Studies on semantic image content in the form of text, face, vehicle, and human action have also attracted some recent interest. 5
  • 6.  Among them, text within an image is of particular interest as (i) It is very useful for describing the contents of an image. (ii)It can be easily extracted compared to other semantic contents, and (iii)It enables applications such as keyword-based image search, and text-based image indexing. 6
  • 7. 7
  • 8. TEXT IN IMAGES  A variety of approaches to Text Information Extraction (TIE) from images have been proposed for specific applications including page segmentation , address block location , license plate location , and content-based image/video indexing  In spite of such extensive studies, it is still not easy to design a general-purpose TIE system. 8
  • 9.  This is because there are so many possible sources of variation when extracting text from a shaded or textured background, from low- contrast or complex images, or from images having variations in font size, style, color , orientation, and alignment.  These variations make the problem of automatic TIE extremely difficult.  Figures 1-4 show some examples of text in images. 9
  • 10.  (a) (b) ©  Fig. 1. Grays cale document images: (a) single- column text from a book,  (b) two-column page from a journal (IEEE Transactions on PAMI), and  (c) an electrical drawing (courtesy of Lu [24]). 10
  • 11. (Fig. 2), Multi-colour document images .  (a) (b) ©  Fig. 2. Multi-color document images: each text line may or may not be of the same color. 11
  • 12. Text in video images can be further classified into caption text (Fig. 3), or scene text (Fig. 4).  Fig.3 (Images with caption text) Text is artificially overlaid on the image. (a) (a) Shows captions overlaid directly on the background 12
  • 13.  Fig. 4. Scene text images: Images with variations in skew, perspective, blur, illumination, and alignment. 13
  • 14. TERMS  Before we attempt to explain the various techniques used in TIE, it is important to define the commonly used terms and summarize the characteristics of text that can be used for TIE algorithms.  Text in images can exhibit many variations with respect to the following properties. 14
  • 15. 1. Geometry  Size: Although the text size can vary a lot, assumptions can be made depending on the application domain.  Alignment: The characters in the caption text appear in clusters and usually lie horizontally.  Although sometimes they can appear as non- planar texts as a result of special effects.  This does not apply to scene text, which can have various perspective distortions. Scene text can be aligned in any direction and can have geometric distortions (Fig. 4). 15
  • 16.  Inter-character distance: characters in a text line have a uniform distance between them. 2. Colour  The characters in a text line tend to have the same or similar colours.  This property makes it possible to use a connected component-based approach for text detection.  Most of the research reported till date has concentrated on finding ‘text strings of a single colour (monochrome)’. 16
  • 17.  However, video images and other complex colour documents can contain ‘text strings with more than two colours (polychrome)’ for effective visualization, i.e., different colours within one word. 3. Motion  The same characters usually exist in consecutive frames in a video with or without movement.  This property is used in text tracking and enhancement.  Caption text usually moves in a uniform way: horizontally or vertically 17
  • 18.  Scene text can have arbitrary motion due to camera or object movement. 4. Edge  Most caption and scene text are designed to be easily read, thereby resulting in strong edges at the boundaries of text and background 5. Compression  Many digital images are recorded, transferred, and processed in a compressed format.  Thus, a faster TIE system can be achieved if one can extract text without decompression 18
  • 19.  The role of text detection is to find the image regions containing only text that can be directly highlighted to the user or fed into an optical character reader module for recognition.  In this seminar a new system is discussing which extracts text in images.  The system takes colored images as input.  It detects text on the basis of certain text features:  The image is then cleaned up so that the text stands out. 19
  • 20. RELATED WORK  Various methods have been proposed in the past for detection and localization of text in images and videos.  These approaches take into consideration different properties related to text in an image such as color, intensity, connected-components, edges etc. 20
  • 21.  These properties are used to distinguish text regions from their background and/or other regions within the image.  In the algorithm based on color clustering, the input image is first pre-processed to remove any noise if present.  Then the image is grouped into different color layers and a gray component.  This approach utilizes the fact that usually the color data in text characters is different from the color data in the background. 21
  • 22.  The potential text regions are localized using connected component based heuristics from these layers.  The experiments conducted show that the algorithm is robust in locating mostly Chinese and English characters in images.  The text detection algorithm is also based on color continuity.  In addition it also uses multi-resolution wavelet transforms and combines low as well as high level image features for text region extraction. 22
  • 23.  Texture based segmentation is used to distinguish text from its background.  Further a bottom-up „chip generation‟ process is carried out which uses the spatial cohesion property of text characters.  The chips are collections of pixels in the image consisting of potential text strokes and edges.  The results show that the algorithm is robust in most cases, except for very small text characters that are not properly detected.  Also in the case of low contrast in the image, misclassifications occur in the texture segmentation. 23
  • 24. METHOD Various steps involved in TIE process are (1) Converting colored image to grayscale (2) Binarization (3) Connected components (4) Horizontal and Vertical Projections (5) Reconstruction. (6) Recognition using OCR 24
  • 25.  Converting colored image to grayscale  A digital color image is a color image that includes color information for each pixel.  There are various color models which are used to represent a color image.  These are RGB color model, in which red, green and blue light is added together in various ways to reproduce a broad array of colors. 25
  • 26.  The other models are CMY color model which uses cyan, magenta and yellow light and HSI model which uses hue, saturation and intensity variations.  Grayscale images have range of shades of gray without apparent color.  These are used as less information needs to be provided for each pixel.  In an 8 bit image, 256 shades are possible.  The darkest possible shade black is represented as 00000000 and lightest possible shade white is represented as 11111111. 26
  • 27.  Binarization  A Binary image is a digital image that can have only two possible values for each pixel.  Each pixel is stored as single bit 0 or 1.  The name black and white is often used for this concept.  To form a binary image we select a threshold intensity value. 27
  • 28.  All the pixels having intensity greater than the threshold value are changes to 0 (black) and the pixels with intensity value less than the threshold intensity value are changed to 1 (white).  Thus the image is changed to a binary image . Connected components  For two pixels to be connected they must be neighbors and their gray levels must specify a certain criterion of similarity. 28
  • 29.  For example, in a binary image with values 0 and 1, two pixels may be neighbors but they are said to be connected only if they have same values.  A pixel p with coordinates (x, y) has four horizontal and vertical neighbors known as 4- neighbors of p, given as: (x, y+1), (x, y-1), (x+1, y), (x-1, y) and four diagonal neighbors given as: (x+1, y-1), (x-1, y-1), (x+1, y+1), (x-1, y+1).  Together these are known as 8-neighbors of p. 29
  • 30.  If S represents subset of pixels in an image, two pixels p and q are said to be connected if there exists a path between them consisting entirely of pixels in S.  For any pixel p in S, the set of pixels that are connected to it in S is called a connected component of S. 30
  • 31. Horizontal and Vertical Projections  The method is performed on binary images.  It starts scanning from left side of every line and records a change in case of facing the pixel change from zero to one and again to zero.  Counting the change does not depend on number of pixels in this method.  Robustness in noisy condition is an advantage of this method.  After finding the connected components, we check the transitions in the values of pixels horizontally. 31
  • 32.  Transitions can be either zero (black) to one (white) or one (white) to zero (black).  The text regions will have larger number of transitions from black to white or vice versa, whereas the background region will have lesser number of transitions.  If the allocated amount of changes for each row is between two thresholds (low and high thresholds), the row potentially would be considered as text area and the up and down of this row would be specified. 32
  • 33.  Next, we search vertically for finding the exact location of the text and ignoring these rows as a text.  For finding the exact location of the text, we use some heuristics.  These heuristics include height and length of the text and the ratio of height to length and enough number of pixels in this area. 33
  • 34. Reconstruction  After the extraction of text regions from images, the text regions become a bit distorted and difficult to read, thus we recover these components using the original image.  The distorted and original images are compared with each other and the pixels which are erased or disfigured are recovered. 34
  • 36. Recognition using OCR  The OCR software converts each character into ASCII codes.  The extracted text is stored as ASCII codes in the computer memory.  OCR systems have been available for a number of years and the current commercial systems can produce an extremely high recognition rate for machine-printed documents on a simple background. 36
  • 37.  Applications  There are numerous applications of a text information extraction system, including document analysis, vehicle license plate extraction, technical paper analysis, and object- oriented data compression.  In the following, we briefly describe some of these applications 37
  • 38.  .  Wearable or portable computers: with the rapid development of computer hardware technology, wearable computers are now a reality.  A TIE system involving a hand-held device and camera was presented as an application of a wearable vision system.  Translation camera can detect text in a scene image and translate Japanese text into English after performing character recognition. 38
  • 39.  Content-based video coding or document coding: The MPEG-4 standard supports object- based encoding.  When text regions are segmented from other regions in an image, this can provide higher compression rates and better image quality.  As a result, they can achieve a higher quality rendering of documents containing text, pictures, and graphics.  Text-based image indexing: This involves automatic text-based video structuring methods using caption data . 39
  • 40.  License/container plate recognition: There has already been a lot of work done on vehicle license plate and container plate recognition.  Although container and vehicle license plates share many characteristics with scene text, many assumptions have been made regarding the image acquisition process (camera and vehicle position and direction, illumination, character types, and color) and geometric attributes of the text. 40
  • 41.  Texts in WWW images: The extraction of text from WWW images can provide relevant information on the Internet.  Video content analysis: Extracted text regions or the output of character recognition can be useful in genre recognition .  The size, position, frequency, text alignment, and OCR-ed results can all be used for this.  Industrial automation: Part identification can be accomplished by using the text information on each part. 41
  • 42. CONCLUSION  Current technologies do not work well for documents with text printed against shaded or textured backgrounds or those with non- structured layout.  In contrast, our technique works well for normal documents as well as with the documents described above situations.  The system is stable and robust. All the system parameters remain the same throughout all the experiments. 42
  • 43.  However, it is not easy to use commercial OCR software for recognizing text extracted from images or video frames.  New OCR systems need to be developed to handle large amount of noise and distortion in TIE applications. 43
  • 44. 44