SlideShare a Scribd company logo
IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 18, Issue 3, Ver. I (May-Jun. 2016), PP 23-29
www.iosrjournals.org
DOI: 10.9790/0661-1803012329 www.iosrjournals.org 23 | Page
Automatic Text Extraction System for Complex Images
R.Janani1
, T.Sheela2
1
(Assistant Professor, Department Of Computer Science and Engineering, Dhanalakshmi Srinivasan
Engineering College, Perambalur-621212, India)
2
(Assistant Professor, Department Of Information Technology, Dhanalakshmi Srinivasan Engineering College,
Perambalur-621212, India)
Abstract : The Intelligent text extraction system automatically identifies and extracts the text present in
different types of images. The growth of digital world Detection and Extraction of text regions in an image are
well known problems in the area of image processing. The growth of digital world and the usage of multimedia
generated a new era with a classic problem of pattern recognition. Thus Automatic text extraction from images
and videos serves an important role in indexing and efficient retrieval of multimedia. The existing techniques
such as region based, texture based techniques for the text extraction are not able to compact with all the
applications of text extraction. The proposed Intelligent Text Extraction system automatically identifies and
extracts the text present in different types of images. The system consists of different stages which include the
localization, segmentation, extraction and recognition of text from the images. In the proposed system we use
the discrete wavelet transform for the localization of text. The morphological operations are used which
enhances the identification of correct text portions. The text part is segmented and is recognized using an
efficient system. The advantage of the system is that the extracted text is shown in the .txt file. The proposed
system also allows the modification of the recognized text from the image. This method shows better efficiency,
precision and recall compared to the existing techniques. This shows the possibility of using this technique in
more new and advanced applications.
Keywords - DWT, image processing, morphological operations, Otsu, segmentation, text extraction, text
recognition
I. Introduction
The rapid advancement in the technology and multimedia has digitalized the world. The availability of
cameras and other systems contributes large number of images to the world. Ranging from cameras embedded
in mobile phones to professional ones, Surveillance cameras to broadcast videos, every day images to satellite
images, all these contributes to increase in multi-media data. Most of the images may contain text as part of it,
which gives some information about that image. Therefore identification of these texts has relevance in many
applications. This shows the importance of the text extraction system in lot of applications. It was stated that in
recent years there was a drastic in-crease in multimedia libraries and the amount of data is growing
exponentially with time. It is also known that there are number of television stations that broad-cast everyday
and the widespread of affordable digital cameras and inexpensive memory devices, the multi-media data is
increasing every second.
The text present within an image enables applications such as keyword-based image search, text-based
image indexing and automatic video logging. Extraction of these texts from images is a very difficult task due to
variations in character fonts, styles, sizes and text directions, and presence of complex backgrounds and variable
light conditions. Generally, the images can be categorized into three based on its type: document images, scene
images and born-digital images. Document images are the image of the document which includes pdf, notes etc.
The text present in document image is the document text.
Fig.1. (a) Document image
Automatic Text Extraction System For Complex Images
DOI: 10.9790/0661-1803012329 www.iosrjournals.org 24 | Page
Fig.1. (b) Scene text image
Fig.1.(a) shows the document image. Born-digital images are the images generated by computer
software. The text in these images is called as caption text or overlay text. Some researchers specify overlay text
as superimposed text or artificial text. An example for overlay text is the text shown by the news channels etc.
Scene images are the images that contain the text, such as the advertising boards, banners, which is
captured naturally when the scene images are taken with the camera as shown in Fig.1.(b), therefore scene text
gets included in the background of image as a part of the scene taken. Compared with document images and
born-digital images, the scene images, have more complex foreground/ background, low resolution,
compression loss, and severe edge softness. This makes the detection of text from scene text more difficult.
There-fore automatic extraction of texts from images or video is a challenging task and research under this field
is still under progress. Text present in images exhibit many variations based on the following properties:
(a) Size: The text can have variable size from small to large.
(b) Alignment: The characters in caption text may appear in clusters and sometimes they can also appear as
non-planar texts. The scene texts possesses numerous perspective distortions .They may be aligned in any
direction with geometric distortions
(c) Inter-character distance: Characters present in a text line have some uniform distance in between them.
(d) Color: In a simple image the characters in a text usually have the similar or same color. Connected
component-based approach usually makes use of this property. But complex images or color documents usually
contain text strings with two or more colors for effective visualization
(e) Edge: Most caption and scene texts are designed to be easily read, therefore strong edges are placed at the
boundaries of text and background.
The remaining part of this paper is organized as follows. Section II describes the related works.
Proposed system is explained in Section III and the experiments and results in section IV. Finally the conclusion
is given in the Section V.
II. Related works
The different approaches related to text extraction includes region based methods, texture based
methods. The region based methods are basically divided into two categories: edge based [1] and connected
component based methods. J.Gllavataet.al [3] proposed a connected component [2] based approach for the text
extraction .It is based on color reduction technique and OCR is used for character recognition .It will only detect
text with horizontal alignment. Low quality images will not be processed accurately. Zhong et al. [4] used a CC-
based method, which uses color reduction. In that they quantize the color space by analyzing the color histogram
generated in the RGB color space. This is done mainly based on the assumption that the text regions usually
cluster together in this color space and they occupy a significant portion of an image. Each text component
present will undergo filtering stage using a number of heuristics, such as area, spatial alignment and diameter.
The performance of this system was evaluated with testing in CD images and other book cover images. Kim et
al. [5] use transition map to detect overlay text. They proposed a method for overlay text detection and
extraction from complex videos. The detection method is based on the observation regarding the existence of
transient color between inserted text and its adjacent background. The transition map is generated first which is
based on its logarithmical change of intensity and modified saturation. Then Linked mask is generated to create
connected components for each candidate region and then each of these connected components is reshaped to
have smooth boundaries. Researches based on Caption text detection are proposed in [9]. Xiaoqing Liu,et.al [6]
proposed a method based on the properties of edges. This method is not sensitive to image color/intensity. Even
though it can handle both printed and document images effectively. Since it mainly analyses texts in the form of
blocks the small image regions and stroke are misidentified as text in areas containing large characters.
Automatic Text Extraction System For Complex Images
DOI: 10.9790/0661-1803012329 www.iosrjournals.org 25 | Page
The proposed system uses wavelet and different morphological operations in a different way compared
to existing techniques for the identification of text part. So a better efficiency is obtained in the extraction of
text. Additional features are also added in our system with the enhancement of technology.
III. Proposed System
The proposed method works based on the fact that the texts present in images have some unique
features which include the properties of edges. The architecture of the proposed system is shown in Fig.2. The
pro-posed system is mainly divided into three modules: Edge map generation module, Text area segmentation
module, Text recognition module. The input is given as image to the system and the output obtained is in the
form of a text file.
1.1 Edge Map Generation Module
The input is passed to this module. The input image can be grayscale or color, compressed or
uncompressed format. This module contains different steps in it. Firstly, the input image undergoes a
preprocessing stage.
Fig.2. Architecture of the proposed system
1.2 Preprocessing
Edge strength and density are fundamental characteristics of that embedded text in a complex image.
This serves as a vital feature for text extraction. In the pro-posed algorithm, the input taken is a color or a gray-
scale image. If the input is a color image it undergoes the preprocessing stage. In this stage, the color image is
converted to grayscale image using the equation below.
(1)
The RGB image is converted to Hue-Saturation-Value (HSV) color space. The Y of the above equation
refers to value component of the Hue-Saturation- Value (HSV) color space. Thus RGB color image gets
converted to the gray scale image. We can also process the image by converting it to YCbCr color space and
taking the Y, the luminance part for processing. On applying a median filtering to this gray scale image the
noise can be filtered out. The best known order statistics filter is the median filter. It replaces the value of a
referred pix-el by another value which is calculated by taking the median of the gray levels in the neighborhood
of that pixel. Median filtering has excellent noise reduction capabilities with less blurring compared to linear
smoothing filters of similar size. It is very effective in case of bipolar and unipolar impulsive noise. After this
filtering step is done, most of noise present in the image gets removed while the edges in the image will be still
preserved. The Image Y is then processed with the 2-D discrete wavelet transform.
3.3 Discrete Wavelet Transform
In this system, the proposed algorithm uses the Haar discrete wavelet trans- form. The Haar DWT
provides a well powerful tool which models most of the characteristics of the image. The textured images are
mostly well characterized by their edges. On applying Discrete Wavelet Transform (DWT), it is decomposed
into components of frequency domain [8]. On applying the DWT the input image is decomposed to four sub-
bands or components. i.e one average component and three detail components. To obtain the components it has
to deal with row and column direction separately. First High Pass Filter (H.P.F) G and Low Pass Filter (L.P.F)
H are exploited for each row data and then are down sampled by 2 to get high and low frequency components of
the row. Next the high and low pass filters are applied again for each high and low frequency components of the
column and then down sampled by 2. By way of the above processing, the four-sub band images are generated:
HH1, HL1, LH1 andLL1.
Automatic Text Extraction System For Complex Images
DOI: 10.9790/0661-1803012329 www.iosrjournals.org 26 | Page
Fig.3. Block diagram of 2D DWT
The detail component sub-bands containing the vertical details LH1, horizontal details HL1 and
diagonal details LL1 are used to detect text edges present in the original image. The process of applying DWT
to image is shown in Fig.3. The D.S. represents the down sampling of the image by 2.
Since filtering is done before applying DWT, the effect of noise in the components can be reduced.
Now edge detection method is applied at each component. The wavelet function and the scaling function of
Haar wavelets are defined below.
(2)
(3)
The Haar is simpler than other wavelets which reduces the complexity of the algorithm. The advantage
of using Haar wavelet is that it is the only wavelet that allows perfect localization in the transform domain. Its
coefficients are either 1 or -1 and are real symmetric and orthogonal. On applying this, the portions with higher
edge strength in identical directions can be found out. Thereafter the threshold is calculated. This helps to filter
out the unfeasible edges. The second derivative of intensity is employed in the measurement of edge strength as
it provides improved detection of intensity places which leads to a characterization of text in images that usually
leads to a characterization of text in images. The traditional edge detection filters may also be able to provide
the similar result but it will not be able to detect three kinds of edges at a time. Therefore, the processing time
for the traditional edge detection filters works slower than 2-D DWT.
3.4 Text Region Detection
The process of identifying text regions can be split into two sub problems: detection and localization.
In the detection step, general regions of the frame are classified as text or non-text. The size and shape of these
regions differ in different algorithm. For example, some algorithms classify 8x8 pixel blocks, while others
classify individual scan lines. In the localization step, the results of detection are grouped together to form one
or more text instances. This is usually represented as a bounding box around each text instance. For the
detection of the text region, the three detailed sub components which is obtained on applying Haar DWT is
used. Sobel edge detection algorithm is applied to each detailed component. Thus hc,vc,dc are images obtained
after edge detection in horizontal ,vertical and diagonal component. Therefore hc contains all the edges in
horizontal direction, vc contains all the edges in vertical direction and the dc contains all the edges in diagonal
direction. Now an edge map is generated using hc, vc and dc. It is generated by using weighted OR operation.
The equation (4) used for the generation of edge map is shown below
I = (40 * hc + 70 * vc + 30 * dc) (4)
The I is the image obtained after applying the edgemap. hc, vc and dc indicates the horizontal, vertical
and diagonal sub component respectively after sobel edge detection algorithm is applied. The edge map of an
example image is shown in Fig 4 .This algorithm uses the sobel edge detector as it is more efficient to locate the
strong edge present. Now the image I shows edges with all the possible text regions.
Automatic Text Extraction System For Complex Images
DOI: 10.9790/0661-1803012329 www.iosrjournals.org 27 | Page
Fig.4. Edgemap of the image obtained
3.5 Text area segmentation
After edge map is obtained, the binarization of the image is done. The binary form of the edge map is
obtained by the thresholding operation. For binarization of the image the Otsu algorithm is used. The
thresholding operation removes the non text regions identified so far. The steps involved in computing Otsu
threshold are
• Reshape the image to 1 dimensional.
• Compute histogram and values at each intensity level
• Initialize a matrix with values from 0 to 255
• Step through all possible thresholds maximum intensity
• Compute the mean, weight and the variance for the foreground and background.
• Calculate weight of the foreground * variance of the foreground + weight of the background * variance of the
background.
• Find the minimum value.
The minimum value calculated is taken as the thresh-old for the binarization process. The localization
process of the text involves further enhancement the text regions by eliminating non-text regions from the
image. One of the main properties the text exhibits is that all characters present will appear close to each other in
the image. Therefore it can form a cluster. By taking into account this property we use morphological
operations. The possible text pixels can be clustered together by using the dilation operation, thus eliminating
pixels that are far from the candidate text regions are possible. Dilation can be defined as an operation that
expands or enhances the region of interest, by the usage of the structural element of the required shape and size.
Large structuring element is used in the dilation process so that regions which lie close to each other can be
enhanced. To localize the text part clearly we use the morphological operations.
The morphological operations include the erosion, dilation with line and disk as the structuring element
is done. The segmentation of the identified text portions are done in this segmentation phase. For that the
connected components are labeled and the connectivity used here is 8. The set of properties, shape and
measurements of the connected components are computed. The area and bounding box shape measurements as
per the requirement are only considered. The area of the connected component is considered as scalar as it
represents the number of pixels in that region.
Bounding box is placed to the connected component identified and it must be the smallest containing
the required text region [11]. The width, height and the up-per left corner position are identified. A new value
can be computed by multiplying height and width of bounding box. The ratio of this new value and area is
taken. If the ratio is less than 1.5, then the regions so obtained are considered as text regions. The resultant
image got after dilation operation may even consist of some non-text regions or any noise which are needed to
be eliminated.
To eliminate noise blobs present in the image an area filtering is done. After that only those regions in
the final image whose area is greater than or equal to 1/15 of the maximum area region identified are retained.
The Fig 5(a) shows the image after morphological operations and Fig 5(b) gives the image after text
extraction.
Fig.5. (a) Image after morphological operations, (b) Output image after segmentation.
Automatic Text Extraction System For Complex Images
DOI: 10.9790/0661-1803012329 www.iosrjournals.org 28 | Page
3.6 Text Recognition
The image got from the previous stage is considered as input in this phase. Text recognition is done by
using the tesseract OCR. The binary image with polygonal text regions defined is given as input to the tesseract.
The processing of the system follows a traditional pipe-line. The first step is analysis of connected component. It
outlines the components stored. This is a computationally expensive design decision, but it has a significant
advantage. At this stage, outlines are gathered together, by nesting, into Blobs. The Blobs thus obtained are
organized into text lines. Then the lines and the regions identified are analyzed for getting fixed pitch or
proportional text. Now the obtained text lines are divided into words in accordance with the character spacing.
Fixed pitch text is then chopped immediately by using the character cells. Proportional text after that is broken
down into words by using definite spaces. Recognition process here undergoes a two-pass process. In the first
pass, an attempt is made to recognize each word. Every word that is satisfactory recognized is then passed to an
adaptive classifier as a training data. The adaptive classifier then will more accurately recognize text in the page.
With the completion of first pass, a second pass is then executed over the page. In this the words that were not
recognized in the previous pass well enough are recognized again. A final phase resolves spaces, and locates the
small text by checking the alternative hypotheses for the x-height to it. After recognition the UTF-8 code of the
character are returned. This can be easily converted corresponding characters and are displayed as output in the
form of a text file.
IV. Experiments and results
The system is experimented with large number of images and the results obtained are described here. It
is clear from the Table 1 that the proposed methods have better precision rate and recall rate compared to other
existing techniques. The method proposed by Samarabandhu et.al shows 91.8 % of precision and 96.6% of
recall whereas the technique given by j.yanget. al have lesser rate of about 84.90% and 90.0 % precision and
recall. But our method shows 99.65% of preci-sion rate and 99.8% of recall. The other existing methods
proposed by Kim et.al and J.Gllavata exhibits even smaller rate as given in the Table 1.
The accuracy of the algorithm in the proposed method is computed by counting the number of correctly
located characters, which is taken as the ground truth. The precision and recall rate are calculated by the
equation (5) & (6)
(5)
(6)
Comparisons with some existing methods are shown in table 1 which shows a clear improvement over
existing methods. The performances of other methods shown are cited from the published works. The proposed
system is implemented using Matlab[13].
Method Precision rate (%) Recall rate (%)
Proposed method 99.65 99.8
Samarabandhu et. al[6] 91.8 96.6
J. yang et. al[11] 84.90 90.0
K.C.kim et al[10] 63.7 82.8
J.Gllavata et.al[3] 83.9 88.7
Table 1 Comparisons with existing methods
Fig.6. The GUI showing the input image and the image with the text portion
Automatic Text Extraction System For Complex Images
DOI: 10.9790/0661-1803012329 www.iosrjournals.org 29 | Page
V. Conclusion
In this paper, a relatively simple, fast and effective method for text detection and extraction are
proposed. The method uses DWT for the efficient working of the algorithm. This process requires less
processing time which is mainly essential for real time applications and shows high precision rate. Most of
methods fail when the characters are not aligned well or when the characters are too small. Those methods also
result in some of missing characters when the characters have very poor contrast with respect to the background.
But the pro-posed method is not sensitive to color or intensity of image, and also the uneven illumination and
reflection effects.
This can be used in large variety of application fields such as vehicle license plate detection, object
identification, identification of various parts in industrial automation, mobile robot navigation which helps to
detect text based land marks, analysis of technical papers with the help of maps, charts, and electric circuits etc.
This algorithm is good at handling both scene text images and documents images effectively. Even though there
are large numbers of algorithms in this area, it is observed that there is no single unified approach or algorithm
that fits for all the applications.
References
[1] Xin Zhang, Fuchun Sun, Lei Gu “A Combined Al-gorithm for Video Text Extraction” Seventh Internation-al Conference on Fuzzy
Systems and Knowledge Discov-ery,2010.
[2] Ohya, A. Shio, and S. Akamatsu, “Recognizing Characters in Scene Images”, IEEE Transactions on Pat-tern Analysis and Machine
Intelligence, 1994, 214-224.
[3] Gllavata, R. Ewerth, and B. Freisleben, “A robust algorithm for text detection in images” Proceedings of the 3rd International
Symposium on Image and Signal Pro-cessing and Analysis, pp.611– 616, ISPA, 2003.
[4] Zhong, Yu.,Karu, K., and Jain, A.K.” Locating text in complex color images” Proceedings of the Third Inter-national Conference
on Document Analysis and Recogni-tion1995, 1, 14-16: 146-149.
[5] WonjunKim,Changick Kim, “A New Approach for Overlay Text Detection and Extraction From Complex Video Scene,” IEEE
Transactions on Image Processing, V.18 , No.2, pp. 401 – 411, 2009.
[6] Xiaoqing Liu, JagathSamarabandu “Multiscale Edge-Based Text Extraction from Complex Images,” International Conference on
Multimedia and Expo, pp.1721-1724, 2006
[7] Xiao-Wei Zhang, Xiong-Bo Zheng, Zhi-Juan Weng, “Text Localization Algorithm Under Back-ground Image Using Wavelet
Transforms”, Proceedings of the 2008 International Conference on Wavelet Analysis and Pattern Recognition, Hong Kong, pp.30-
31, Aug. 2008.
[8] N. Otsu, “A Threshold Selection Method from Gray-Level Histograms,” IEEE Transactions on Systems, man and Cybernet, 1979.
[9] Debapratim Sarkar, Raghunath Ghosh ,”A Bottom-Up Approach of Line Segmentation from Handwritten Text”2009.
[10] K.C. Kim, H.R. Byun, Y.J. Song, Y.W. Choi, S.Y. Chi,K.K. Kim and Y.K Chung, Scene Text Extraction in Natural Scene Images
using Hierarchical FeatureCom-bining and verification, Proceedings of the 17International Conference on Pattern Recognition
(ICPR’04),IEEE.
[11] J. Yang, J. Gao, Y. Zhang, X. Chen and A. Waibel, “AnAutomatic Sign Recognition and Translation Sys-tem”,Proceedings of the
Workshop on Perceptive User Inter-faces(PUI'01), 2001, pp. 1-8.
[12] A.K Jain,”Fundamentals of Digital Image Pro-cessing”,Englewood cliff, NJ: Prentice Hall, 1989.

More Related Content

What's hot

26 3 jul17 22may 6664 8052-1-ed edit septian
26 3 jul17 22may 6664 8052-1-ed edit septian26 3 jul17 22may 6664 8052-1-ed edit septian
26 3 jul17 22may 6664 8052-1-ed edit septian
IAESIJEECS
 
Document Recovery From Degraded Images
Document Recovery From Degraded ImagesDocument Recovery From Degraded Images
Document Recovery From Degraded Images
IRJET Journal
 
Block Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid ExtensionBlock Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid Extension
DR.P.S.JAGADEESH KUMAR
 
Scene Text Detection of Curved Text Using Gradiant Vector Flow Method
Scene Text Detection of Curved Text Using Gradiant Vector Flow MethodScene Text Detection of Curved Text Using Gradiant Vector Flow Method
Scene Text Detection of Curved Text Using Gradiant Vector Flow Method
IJTET Journal
 
2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...
2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...
2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...
INFOGAIN PUBLICATION
 
A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In Videos
CSCJournals
 
Modified Skip Line Encoding for Binary Image Compression
Modified Skip Line Encoding for Binary Image CompressionModified Skip Line Encoding for Binary Image Compression
Modified Skip Line Encoding for Binary Image Compression
idescitation
 
IRJET-MText Extraction from Images using Convolutional Neural Network
IRJET-MText Extraction from Images using Convolutional Neural NetworkIRJET-MText Extraction from Images using Convolutional Neural Network
IRJET-MText Extraction from Images using Convolutional Neural Network
IRJET Journal
 
A Study of Image Compression Methods
A Study of Image Compression MethodsA Study of Image Compression Methods
A Study of Image Compression Methods
IOSR Journals
 
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
IJERA Editor
 
A Novel Document Image Binarization For Optical Character Recognition
A Novel Document Image Binarization For Optical Character RecognitionA Novel Document Image Binarization For Optical Character Recognition
A Novel Document Image Binarization For Optical Character Recognition
Editor IJCATR
 
Embedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTPEmbedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTP
csijjournal
 
Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...
Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...
Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...
cscpconf
 
Facial image retrieval on semantic features using adaptive mean genetic algor...
Facial image retrieval on semantic features using adaptive mean genetic algor...Facial image retrieval on semantic features using adaptive mean genetic algor...
Facial image retrieval on semantic features using adaptive mean genetic algor...
TELKOMNIKA JOURNAL
 
MMFO: modified moth flame optimization algorithm for region based RGB color i...
MMFO: modified moth flame optimization algorithm for region based RGB color i...MMFO: modified moth flame optimization algorithm for region based RGB color i...
MMFO: modified moth flame optimization algorithm for region based RGB color i...
IJECEIAES
 
Wavelet-Based Color Histogram on Content-Based Image Retrieval
Wavelet-Based Color Histogram on Content-Based Image RetrievalWavelet-Based Color Histogram on Content-Based Image Retrieval
Wavelet-Based Color Histogram on Content-Based Image Retrieval
TELKOMNIKA JOURNAL
 

What's hot (16)

26 3 jul17 22may 6664 8052-1-ed edit septian
26 3 jul17 22may 6664 8052-1-ed edit septian26 3 jul17 22may 6664 8052-1-ed edit septian
26 3 jul17 22may 6664 8052-1-ed edit septian
 
Document Recovery From Degraded Images
Document Recovery From Degraded ImagesDocument Recovery From Degraded Images
Document Recovery From Degraded Images
 
Block Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid ExtensionBlock Classification Scheme of Compound Images: A Hybrid Extension
Block Classification Scheme of Compound Images: A Hybrid Extension
 
Scene Text Detection of Curved Text Using Gradiant Vector Flow Method
Scene Text Detection of Curved Text Using Gradiant Vector Flow MethodScene Text Detection of Curved Text Using Gradiant Vector Flow Method
Scene Text Detection of Curved Text Using Gradiant Vector Flow Method
 
2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...
2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...
2 ijaems dec-2015-5-comprehensive review of huffman encoding technique for im...
 
A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In Videos
 
Modified Skip Line Encoding for Binary Image Compression
Modified Skip Line Encoding for Binary Image CompressionModified Skip Line Encoding for Binary Image Compression
Modified Skip Line Encoding for Binary Image Compression
 
IRJET-MText Extraction from Images using Convolutional Neural Network
IRJET-MText Extraction from Images using Convolutional Neural NetworkIRJET-MText Extraction from Images using Convolutional Neural Network
IRJET-MText Extraction from Images using Convolutional Neural Network
 
A Study of Image Compression Methods
A Study of Image Compression MethodsA Study of Image Compression Methods
A Study of Image Compression Methods
 
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
Query Image Searching With Integrated Textual and Visual Relevance Feedback f...
 
A Novel Document Image Binarization For Optical Character Recognition
A Novel Document Image Binarization For Optical Character RecognitionA Novel Document Image Binarization For Optical Character Recognition
A Novel Document Image Binarization For Optical Character Recognition
 
Embedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTPEmbedding Patient Information In Medical Images Using LBP and LTP
Embedding Patient Information In Medical Images Using LBP and LTP
 
Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...
Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...
Content Based Image Retrieval Approach Based on Top-Hat Transform And Modifie...
 
Facial image retrieval on semantic features using adaptive mean genetic algor...
Facial image retrieval on semantic features using adaptive mean genetic algor...Facial image retrieval on semantic features using adaptive mean genetic algor...
Facial image retrieval on semantic features using adaptive mean genetic algor...
 
MMFO: modified moth flame optimization algorithm for region based RGB color i...
MMFO: modified moth flame optimization algorithm for region based RGB color i...MMFO: modified moth flame optimization algorithm for region based RGB color i...
MMFO: modified moth flame optimization algorithm for region based RGB color i...
 
Wavelet-Based Color Histogram on Content-Based Image Retrieval
Wavelet-Based Color Histogram on Content-Based Image RetrievalWavelet-Based Color Histogram on Content-Based Image Retrieval
Wavelet-Based Color Histogram on Content-Based Image Retrieval
 

Viewers also liked

B017221317
B017221317B017221317
B017221317
IOSR Journals
 
O130303104110
O130303104110O130303104110
O130303104110
IOSR Journals
 
I012274853
I012274853I012274853
I012274853
IOSR Journals
 
O1304039196
O1304039196O1304039196
O1304039196
IOSR Journals
 
E1303041923
E1303041923E1303041923
E1303041923
IOSR Journals
 
B012620814
B012620814B012620814
B012620814
IOSR Journals
 
F010133036
F010133036F010133036
F010133036
IOSR Journals
 
Design and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEA
Design and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEADesign and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEA
Design and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEA
IOSR Journals
 
Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...
Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...
Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...
IOSR Journals
 
Multiplier and Accumulator Using Csla
Multiplier and Accumulator Using CslaMultiplier and Accumulator Using Csla
Multiplier and Accumulator Using Csla
IOSR Journals
 
D012241727
D012241727D012241727
D012241727
IOSR Journals
 
Energy Policy in India
Energy Policy in IndiaEnergy Policy in India
Energy Policy in India
IOSR Journals
 
R01821116120
R01821116120R01821116120
R01821116120
IOSR Journals
 
F0552935
F0552935F0552935
F0552935
IOSR Journals
 
I012135157
I012135157I012135157
I012135157
IOSR Journals
 
Feature Based Semantic Polarity Analysis Through Ontology
Feature Based Semantic Polarity Analysis Through OntologyFeature Based Semantic Polarity Analysis Through Ontology
Feature Based Semantic Polarity Analysis Through Ontology
IOSR Journals
 
Application of Roller Burnishing Process for Final Machining Of Cilindrical S...
Application of Roller Burnishing Process for Final Machining Of Cilindrical S...Application of Roller Burnishing Process for Final Machining Of Cilindrical S...
Application of Roller Burnishing Process for Final Machining Of Cilindrical S...
IOSR Journals
 
Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...
Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...
Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...
IOSR Journals
 
J012455865
J012455865J012455865
J012455865
IOSR Journals
 
E010342327
E010342327E010342327
E010342327
IOSR Journals
 

Viewers also liked (20)

B017221317
B017221317B017221317
B017221317
 
O130303104110
O130303104110O130303104110
O130303104110
 
I012274853
I012274853I012274853
I012274853
 
O1304039196
O1304039196O1304039196
O1304039196
 
E1303041923
E1303041923E1303041923
E1303041923
 
B012620814
B012620814B012620814
B012620814
 
F010133036
F010133036F010133036
F010133036
 
Design and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEA
Design and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEADesign and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEA
Design and Analysis of Mono Composite Leaf Spring by Varying Thickness using FEA
 
Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...
Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...
Evaluation of Radiation Hazard Indices and Excess Lifetime Cancer Risk Due To...
 
Multiplier and Accumulator Using Csla
Multiplier and Accumulator Using CslaMultiplier and Accumulator Using Csla
Multiplier and Accumulator Using Csla
 
D012241727
D012241727D012241727
D012241727
 
Energy Policy in India
Energy Policy in IndiaEnergy Policy in India
Energy Policy in India
 
R01821116120
R01821116120R01821116120
R01821116120
 
F0552935
F0552935F0552935
F0552935
 
I012135157
I012135157I012135157
I012135157
 
Feature Based Semantic Polarity Analysis Through Ontology
Feature Based Semantic Polarity Analysis Through OntologyFeature Based Semantic Polarity Analysis Through Ontology
Feature Based Semantic Polarity Analysis Through Ontology
 
Application of Roller Burnishing Process for Final Machining Of Cilindrical S...
Application of Roller Burnishing Process for Final Machining Of Cilindrical S...Application of Roller Burnishing Process for Final Machining Of Cilindrical S...
Application of Roller Burnishing Process for Final Machining Of Cilindrical S...
 
Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...
Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...
Crack Detection for Various Loading Conditions in Beam Using Hilbert – Huang ...
 
J012455865
J012455865J012455865
J012455865
 
E010342327
E010342327E010342327
E010342327
 

Similar to E1803012329

A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In Videos
CSCJournals
 
Text extraction from images
Text extraction from imagesText extraction from images
Text extraction from images
Garby Baby
 
On Text Realization Image Steganography
On Text Realization Image SteganographyOn Text Realization Image Steganography
On Text Realization Image Steganography
CSCJournals
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Editor IJARCET
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Editor IJARCET
 
Inpainting scheme for text in video a survey
Inpainting scheme for text in video   a surveyInpainting scheme for text in video   a survey
Inpainting scheme for text in video a survey
eSAT Journals
 
Comparative analysis of c99 and topictiling text segmentation algorithms
Comparative analysis of c99 and topictiling text segmentation algorithmsComparative analysis of c99 and topictiling text segmentation algorithms
Comparative analysis of c99 and topictiling text segmentation algorithms
eSAT Journals
 
Comparative analysis of c99 and topictiling text
Comparative analysis of c99 and topictiling textComparative analysis of c99 and topictiling text
Comparative analysis of c99 and topictiling text
eSAT Publishing House
 
Fc4301935938
Fc4301935938Fc4301935938
Fc4301935938
IJERA Editor
 
AVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification SchemeAVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification Scheme
DR.P.S.JAGADEESH KUMAR
 
A Texture Based Methodology for Text Region Extraction from Low Resolution Na...
A Texture Based Methodology for Text Region Extraction from Low Resolution Na...A Texture Based Methodology for Text Region Extraction from Low Resolution Na...
A Texture Based Methodology for Text Region Extraction from Low Resolution Na...
CSCJournals
 
C04741319
C04741319C04741319
C04741319
IOSR-JEN
 
F045053236
F045053236F045053236
F045053236
IJERA Editor
 
E017322833
E017322833E017322833
E017322833
IOSR Journals
 
A Review of Optical Character Recognition System for Recognition of Printed Text
A Review of Optical Character Recognition System for Recognition of Printed TextA Review of Optical Character Recognition System for Recognition of Printed Text
A Review of Optical Character Recognition System for Recognition of Printed Text
iosrjce
 
A comparative study on content based image retrieval methods
A comparative study on content based image retrieval methodsA comparative study on content based image retrieval methods
A comparative study on content based image retrieval methods
IJLT EMAS
 
IMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNING
IMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNINGIMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNING
IMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNING
IRJET Journal
 
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIESLOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
aciijournal
 
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIESLOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIESaciijournal
 

Similar to E1803012329 (20)

A Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In VideosA Survey On Thresholding Operators of Text Extraction In Videos
A Survey On Thresholding Operators of Text Extraction In Videos
 
Text extraction from images
Text extraction from imagesText extraction from images
Text extraction from images
 
On Text Realization Image Steganography
On Text Realization Image SteganographyOn Text Realization Image Steganography
On Text Realization Image Steganography
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124
 
Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124Volume 2-issue-6-2119-2124
Volume 2-issue-6-2119-2124
 
Inpainting scheme for text in video a survey
Inpainting scheme for text in video   a surveyInpainting scheme for text in video   a survey
Inpainting scheme for text in video a survey
 
Comparative analysis of c99 and topictiling text segmentation algorithms
Comparative analysis of c99 and topictiling text segmentation algorithmsComparative analysis of c99 and topictiling text segmentation algorithms
Comparative analysis of c99 and topictiling text segmentation algorithms
 
Comparative analysis of c99 and topictiling text
Comparative analysis of c99 and topictiling textComparative analysis of c99 and topictiling text
Comparative analysis of c99 and topictiling text
 
Fc4301935938
Fc4301935938Fc4301935938
Fc4301935938
 
AVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification SchemeAVC based Compression of Compound Images Using Block Classification Scheme
AVC based Compression of Compound Images Using Block Classification Scheme
 
A Texture Based Methodology for Text Region Extraction from Low Resolution Na...
A Texture Based Methodology for Text Region Extraction from Low Resolution Na...A Texture Based Methodology for Text Region Extraction from Low Resolution Na...
A Texture Based Methodology for Text Region Extraction from Low Resolution Na...
 
C04741319
C04741319C04741319
C04741319
 
F045053236
F045053236F045053236
F045053236
 
40120140501009
4012014050100940120140501009
40120140501009
 
E017322833
E017322833E017322833
E017322833
 
A Review of Optical Character Recognition System for Recognition of Printed Text
A Review of Optical Character Recognition System for Recognition of Printed TextA Review of Optical Character Recognition System for Recognition of Printed Text
A Review of Optical Character Recognition System for Recognition of Printed Text
 
A comparative study on content based image retrieval methods
A comparative study on content based image retrieval methodsA comparative study on content based image retrieval methods
A comparative study on content based image retrieval methods
 
IMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNING
IMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNINGIMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNING
IMAGE TO TEXT TO SPEECH CONVERSION USING MACHINE LEARNING
 
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIESLOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
 
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIESLOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
LOCALIZATION OF OVERLAID TEXT BASED ON NOISE INCONSISTENCIES
 

More from IOSR Journals

A011140104
A011140104A011140104
A011140104
IOSR Journals
 
M0111397100
M0111397100M0111397100
M0111397100
IOSR Journals
 
L011138596
L011138596L011138596
L011138596
IOSR Journals
 
K011138084
K011138084K011138084
K011138084
IOSR Journals
 
J011137479
J011137479J011137479
J011137479
IOSR Journals
 
I011136673
I011136673I011136673
I011136673
IOSR Journals
 
G011134454
G011134454G011134454
G011134454
IOSR Journals
 
H011135565
H011135565H011135565
H011135565
IOSR Journals
 
F011134043
F011134043F011134043
F011134043
IOSR Journals
 
E011133639
E011133639E011133639
E011133639
IOSR Journals
 
D011132635
D011132635D011132635
D011132635
IOSR Journals
 
C011131925
C011131925C011131925
C011131925
IOSR Journals
 
B011130918
B011130918B011130918
B011130918
IOSR Journals
 
A011130108
A011130108A011130108
A011130108
IOSR Journals
 
I011125160
I011125160I011125160
I011125160
IOSR Journals
 
H011124050
H011124050H011124050
H011124050
IOSR Journals
 
G011123539
G011123539G011123539
G011123539
IOSR Journals
 
F011123134
F011123134F011123134
F011123134
IOSR Journals
 
E011122530
E011122530E011122530
E011122530
IOSR Journals
 
D011121524
D011121524D011121524
D011121524
IOSR Journals
 

More from IOSR Journals (20)

A011140104
A011140104A011140104
A011140104
 
M0111397100
M0111397100M0111397100
M0111397100
 
L011138596
L011138596L011138596
L011138596
 
K011138084
K011138084K011138084
K011138084
 
J011137479
J011137479J011137479
J011137479
 
I011136673
I011136673I011136673
I011136673
 
G011134454
G011134454G011134454
G011134454
 
H011135565
H011135565H011135565
H011135565
 
F011134043
F011134043F011134043
F011134043
 
E011133639
E011133639E011133639
E011133639
 
D011132635
D011132635D011132635
D011132635
 
C011131925
C011131925C011131925
C011131925
 
B011130918
B011130918B011130918
B011130918
 
A011130108
A011130108A011130108
A011130108
 
I011125160
I011125160I011125160
I011125160
 
H011124050
H011124050H011124050
H011124050
 
G011123539
G011123539G011123539
G011123539
 
F011123134
F011123134F011123134
F011123134
 
E011122530
E011122530E011122530
E011122530
 
D011121524
D011121524D011121524
D011121524
 

Recently uploaded

Key Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdfKey Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdf
Cheryl Hung
 
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
UiPathCommunity
 
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Jeffrey Haguewood
 
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdfFIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
FIDO Alliance
 
When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...
Elena Simperl
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
OnBoard
 
The Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and SalesThe Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and Sales
Laura Byrne
 
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
Product School
 
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdfFIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance
 
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Thierry Lestable
 
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdfFIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance
 
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Tobias Schneck
 
Connector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a buttonConnector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a button
DianaGray10
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
Alan Dix
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and back
Elena Simperl
 
Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...
Product School
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
Guy Korland
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
Jemma Hussein Allen
 
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdfSmart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
91mobiles
 
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
Product School
 

Recently uploaded (20)

Key Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdfKey Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdf
 
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...
 
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
 
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdfFIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
FIDO Alliance Osaka Seminar: Passkeys and the Road Ahead.pdf
 
When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
 
The Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and SalesThe Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and Sales
 
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
Unsubscribed: Combat Subscription Fatigue With a Membership Mentality by Head...
 
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdfFIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdf
 
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
Empowering NextGen Mobility via Large Action Model Infrastructure (LAMI): pav...
 
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdfFIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
FIDO Alliance Osaka Seminar: The WebAuthn API and Discoverable Credentials.pdf
 
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
 
Connector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a buttonConnector Corner: Automate dynamic content and events by pushing a button
Connector Corner: Automate dynamic content and events by pushing a button
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and back
 
Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...Designing Great Products: The Power of Design and Leadership by Chief Designe...
Designing Great Products: The Power of Design and Leadership by Chief Designe...
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
 
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdfSmart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
 
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
AI for Every Business: Unlocking Your Product's Universal Potential by VP of ...
 

E1803012329

  • 1. IOSR Journal of Computer Engineering (IOSR-JCE) e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 18, Issue 3, Ver. I (May-Jun. 2016), PP 23-29 www.iosrjournals.org DOI: 10.9790/0661-1803012329 www.iosrjournals.org 23 | Page Automatic Text Extraction System for Complex Images R.Janani1 , T.Sheela2 1 (Assistant Professor, Department Of Computer Science and Engineering, Dhanalakshmi Srinivasan Engineering College, Perambalur-621212, India) 2 (Assistant Professor, Department Of Information Technology, Dhanalakshmi Srinivasan Engineering College, Perambalur-621212, India) Abstract : The Intelligent text extraction system automatically identifies and extracts the text present in different types of images. The growth of digital world Detection and Extraction of text regions in an image are well known problems in the area of image processing. The growth of digital world and the usage of multimedia generated a new era with a classic problem of pattern recognition. Thus Automatic text extraction from images and videos serves an important role in indexing and efficient retrieval of multimedia. The existing techniques such as region based, texture based techniques for the text extraction are not able to compact with all the applications of text extraction. The proposed Intelligent Text Extraction system automatically identifies and extracts the text present in different types of images. The system consists of different stages which include the localization, segmentation, extraction and recognition of text from the images. In the proposed system we use the discrete wavelet transform for the localization of text. The morphological operations are used which enhances the identification of correct text portions. The text part is segmented and is recognized using an efficient system. The advantage of the system is that the extracted text is shown in the .txt file. The proposed system also allows the modification of the recognized text from the image. This method shows better efficiency, precision and recall compared to the existing techniques. This shows the possibility of using this technique in more new and advanced applications. Keywords - DWT, image processing, morphological operations, Otsu, segmentation, text extraction, text recognition I. Introduction The rapid advancement in the technology and multimedia has digitalized the world. The availability of cameras and other systems contributes large number of images to the world. Ranging from cameras embedded in mobile phones to professional ones, Surveillance cameras to broadcast videos, every day images to satellite images, all these contributes to increase in multi-media data. Most of the images may contain text as part of it, which gives some information about that image. Therefore identification of these texts has relevance in many applications. This shows the importance of the text extraction system in lot of applications. It was stated that in recent years there was a drastic in-crease in multimedia libraries and the amount of data is growing exponentially with time. It is also known that there are number of television stations that broad-cast everyday and the widespread of affordable digital cameras and inexpensive memory devices, the multi-media data is increasing every second. The text present within an image enables applications such as keyword-based image search, text-based image indexing and automatic video logging. Extraction of these texts from images is a very difficult task due to variations in character fonts, styles, sizes and text directions, and presence of complex backgrounds and variable light conditions. Generally, the images can be categorized into three based on its type: document images, scene images and born-digital images. Document images are the image of the document which includes pdf, notes etc. The text present in document image is the document text. Fig.1. (a) Document image
  • 2. Automatic Text Extraction System For Complex Images DOI: 10.9790/0661-1803012329 www.iosrjournals.org 24 | Page Fig.1. (b) Scene text image Fig.1.(a) shows the document image. Born-digital images are the images generated by computer software. The text in these images is called as caption text or overlay text. Some researchers specify overlay text as superimposed text or artificial text. An example for overlay text is the text shown by the news channels etc. Scene images are the images that contain the text, such as the advertising boards, banners, which is captured naturally when the scene images are taken with the camera as shown in Fig.1.(b), therefore scene text gets included in the background of image as a part of the scene taken. Compared with document images and born-digital images, the scene images, have more complex foreground/ background, low resolution, compression loss, and severe edge softness. This makes the detection of text from scene text more difficult. There-fore automatic extraction of texts from images or video is a challenging task and research under this field is still under progress. Text present in images exhibit many variations based on the following properties: (a) Size: The text can have variable size from small to large. (b) Alignment: The characters in caption text may appear in clusters and sometimes they can also appear as non-planar texts. The scene texts possesses numerous perspective distortions .They may be aligned in any direction with geometric distortions (c) Inter-character distance: Characters present in a text line have some uniform distance in between them. (d) Color: In a simple image the characters in a text usually have the similar or same color. Connected component-based approach usually makes use of this property. But complex images or color documents usually contain text strings with two or more colors for effective visualization (e) Edge: Most caption and scene texts are designed to be easily read, therefore strong edges are placed at the boundaries of text and background. The remaining part of this paper is organized as follows. Section II describes the related works. Proposed system is explained in Section III and the experiments and results in section IV. Finally the conclusion is given in the Section V. II. Related works The different approaches related to text extraction includes region based methods, texture based methods. The region based methods are basically divided into two categories: edge based [1] and connected component based methods. J.Gllavataet.al [3] proposed a connected component [2] based approach for the text extraction .It is based on color reduction technique and OCR is used for character recognition .It will only detect text with horizontal alignment. Low quality images will not be processed accurately. Zhong et al. [4] used a CC- based method, which uses color reduction. In that they quantize the color space by analyzing the color histogram generated in the RGB color space. This is done mainly based on the assumption that the text regions usually cluster together in this color space and they occupy a significant portion of an image. Each text component present will undergo filtering stage using a number of heuristics, such as area, spatial alignment and diameter. The performance of this system was evaluated with testing in CD images and other book cover images. Kim et al. [5] use transition map to detect overlay text. They proposed a method for overlay text detection and extraction from complex videos. The detection method is based on the observation regarding the existence of transient color between inserted text and its adjacent background. The transition map is generated first which is based on its logarithmical change of intensity and modified saturation. Then Linked mask is generated to create connected components for each candidate region and then each of these connected components is reshaped to have smooth boundaries. Researches based on Caption text detection are proposed in [9]. Xiaoqing Liu,et.al [6] proposed a method based on the properties of edges. This method is not sensitive to image color/intensity. Even though it can handle both printed and document images effectively. Since it mainly analyses texts in the form of blocks the small image regions and stroke are misidentified as text in areas containing large characters.
  • 3. Automatic Text Extraction System For Complex Images DOI: 10.9790/0661-1803012329 www.iosrjournals.org 25 | Page The proposed system uses wavelet and different morphological operations in a different way compared to existing techniques for the identification of text part. So a better efficiency is obtained in the extraction of text. Additional features are also added in our system with the enhancement of technology. III. Proposed System The proposed method works based on the fact that the texts present in images have some unique features which include the properties of edges. The architecture of the proposed system is shown in Fig.2. The pro-posed system is mainly divided into three modules: Edge map generation module, Text area segmentation module, Text recognition module. The input is given as image to the system and the output obtained is in the form of a text file. 1.1 Edge Map Generation Module The input is passed to this module. The input image can be grayscale or color, compressed or uncompressed format. This module contains different steps in it. Firstly, the input image undergoes a preprocessing stage. Fig.2. Architecture of the proposed system 1.2 Preprocessing Edge strength and density are fundamental characteristics of that embedded text in a complex image. This serves as a vital feature for text extraction. In the pro-posed algorithm, the input taken is a color or a gray- scale image. If the input is a color image it undergoes the preprocessing stage. In this stage, the color image is converted to grayscale image using the equation below. (1) The RGB image is converted to Hue-Saturation-Value (HSV) color space. The Y of the above equation refers to value component of the Hue-Saturation- Value (HSV) color space. Thus RGB color image gets converted to the gray scale image. We can also process the image by converting it to YCbCr color space and taking the Y, the luminance part for processing. On applying a median filtering to this gray scale image the noise can be filtered out. The best known order statistics filter is the median filter. It replaces the value of a referred pix-el by another value which is calculated by taking the median of the gray levels in the neighborhood of that pixel. Median filtering has excellent noise reduction capabilities with less blurring compared to linear smoothing filters of similar size. It is very effective in case of bipolar and unipolar impulsive noise. After this filtering step is done, most of noise present in the image gets removed while the edges in the image will be still preserved. The Image Y is then processed with the 2-D discrete wavelet transform. 3.3 Discrete Wavelet Transform In this system, the proposed algorithm uses the Haar discrete wavelet trans- form. The Haar DWT provides a well powerful tool which models most of the characteristics of the image. The textured images are mostly well characterized by their edges. On applying Discrete Wavelet Transform (DWT), it is decomposed into components of frequency domain [8]. On applying the DWT the input image is decomposed to four sub- bands or components. i.e one average component and three detail components. To obtain the components it has to deal with row and column direction separately. First High Pass Filter (H.P.F) G and Low Pass Filter (L.P.F) H are exploited for each row data and then are down sampled by 2 to get high and low frequency components of the row. Next the high and low pass filters are applied again for each high and low frequency components of the column and then down sampled by 2. By way of the above processing, the four-sub band images are generated: HH1, HL1, LH1 andLL1.
  • 4. Automatic Text Extraction System For Complex Images DOI: 10.9790/0661-1803012329 www.iosrjournals.org 26 | Page Fig.3. Block diagram of 2D DWT The detail component sub-bands containing the vertical details LH1, horizontal details HL1 and diagonal details LL1 are used to detect text edges present in the original image. The process of applying DWT to image is shown in Fig.3. The D.S. represents the down sampling of the image by 2. Since filtering is done before applying DWT, the effect of noise in the components can be reduced. Now edge detection method is applied at each component. The wavelet function and the scaling function of Haar wavelets are defined below. (2) (3) The Haar is simpler than other wavelets which reduces the complexity of the algorithm. The advantage of using Haar wavelet is that it is the only wavelet that allows perfect localization in the transform domain. Its coefficients are either 1 or -1 and are real symmetric and orthogonal. On applying this, the portions with higher edge strength in identical directions can be found out. Thereafter the threshold is calculated. This helps to filter out the unfeasible edges. The second derivative of intensity is employed in the measurement of edge strength as it provides improved detection of intensity places which leads to a characterization of text in images that usually leads to a characterization of text in images. The traditional edge detection filters may also be able to provide the similar result but it will not be able to detect three kinds of edges at a time. Therefore, the processing time for the traditional edge detection filters works slower than 2-D DWT. 3.4 Text Region Detection The process of identifying text regions can be split into two sub problems: detection and localization. In the detection step, general regions of the frame are classified as text or non-text. The size and shape of these regions differ in different algorithm. For example, some algorithms classify 8x8 pixel blocks, while others classify individual scan lines. In the localization step, the results of detection are grouped together to form one or more text instances. This is usually represented as a bounding box around each text instance. For the detection of the text region, the three detailed sub components which is obtained on applying Haar DWT is used. Sobel edge detection algorithm is applied to each detailed component. Thus hc,vc,dc are images obtained after edge detection in horizontal ,vertical and diagonal component. Therefore hc contains all the edges in horizontal direction, vc contains all the edges in vertical direction and the dc contains all the edges in diagonal direction. Now an edge map is generated using hc, vc and dc. It is generated by using weighted OR operation. The equation (4) used for the generation of edge map is shown below I = (40 * hc + 70 * vc + 30 * dc) (4) The I is the image obtained after applying the edgemap. hc, vc and dc indicates the horizontal, vertical and diagonal sub component respectively after sobel edge detection algorithm is applied. The edge map of an example image is shown in Fig 4 .This algorithm uses the sobel edge detector as it is more efficient to locate the strong edge present. Now the image I shows edges with all the possible text regions.
  • 5. Automatic Text Extraction System For Complex Images DOI: 10.9790/0661-1803012329 www.iosrjournals.org 27 | Page Fig.4. Edgemap of the image obtained 3.5 Text area segmentation After edge map is obtained, the binarization of the image is done. The binary form of the edge map is obtained by the thresholding operation. For binarization of the image the Otsu algorithm is used. The thresholding operation removes the non text regions identified so far. The steps involved in computing Otsu threshold are • Reshape the image to 1 dimensional. • Compute histogram and values at each intensity level • Initialize a matrix with values from 0 to 255 • Step through all possible thresholds maximum intensity • Compute the mean, weight and the variance for the foreground and background. • Calculate weight of the foreground * variance of the foreground + weight of the background * variance of the background. • Find the minimum value. The minimum value calculated is taken as the thresh-old for the binarization process. The localization process of the text involves further enhancement the text regions by eliminating non-text regions from the image. One of the main properties the text exhibits is that all characters present will appear close to each other in the image. Therefore it can form a cluster. By taking into account this property we use morphological operations. The possible text pixels can be clustered together by using the dilation operation, thus eliminating pixels that are far from the candidate text regions are possible. Dilation can be defined as an operation that expands or enhances the region of interest, by the usage of the structural element of the required shape and size. Large structuring element is used in the dilation process so that regions which lie close to each other can be enhanced. To localize the text part clearly we use the morphological operations. The morphological operations include the erosion, dilation with line and disk as the structuring element is done. The segmentation of the identified text portions are done in this segmentation phase. For that the connected components are labeled and the connectivity used here is 8. The set of properties, shape and measurements of the connected components are computed. The area and bounding box shape measurements as per the requirement are only considered. The area of the connected component is considered as scalar as it represents the number of pixels in that region. Bounding box is placed to the connected component identified and it must be the smallest containing the required text region [11]. The width, height and the up-per left corner position are identified. A new value can be computed by multiplying height and width of bounding box. The ratio of this new value and area is taken. If the ratio is less than 1.5, then the regions so obtained are considered as text regions. The resultant image got after dilation operation may even consist of some non-text regions or any noise which are needed to be eliminated. To eliminate noise blobs present in the image an area filtering is done. After that only those regions in the final image whose area is greater than or equal to 1/15 of the maximum area region identified are retained. The Fig 5(a) shows the image after morphological operations and Fig 5(b) gives the image after text extraction. Fig.5. (a) Image after morphological operations, (b) Output image after segmentation.
  • 6. Automatic Text Extraction System For Complex Images DOI: 10.9790/0661-1803012329 www.iosrjournals.org 28 | Page 3.6 Text Recognition The image got from the previous stage is considered as input in this phase. Text recognition is done by using the tesseract OCR. The binary image with polygonal text regions defined is given as input to the tesseract. The processing of the system follows a traditional pipe-line. The first step is analysis of connected component. It outlines the components stored. This is a computationally expensive design decision, but it has a significant advantage. At this stage, outlines are gathered together, by nesting, into Blobs. The Blobs thus obtained are organized into text lines. Then the lines and the regions identified are analyzed for getting fixed pitch or proportional text. Now the obtained text lines are divided into words in accordance with the character spacing. Fixed pitch text is then chopped immediately by using the character cells. Proportional text after that is broken down into words by using definite spaces. Recognition process here undergoes a two-pass process. In the first pass, an attempt is made to recognize each word. Every word that is satisfactory recognized is then passed to an adaptive classifier as a training data. The adaptive classifier then will more accurately recognize text in the page. With the completion of first pass, a second pass is then executed over the page. In this the words that were not recognized in the previous pass well enough are recognized again. A final phase resolves spaces, and locates the small text by checking the alternative hypotheses for the x-height to it. After recognition the UTF-8 code of the character are returned. This can be easily converted corresponding characters and are displayed as output in the form of a text file. IV. Experiments and results The system is experimented with large number of images and the results obtained are described here. It is clear from the Table 1 that the proposed methods have better precision rate and recall rate compared to other existing techniques. The method proposed by Samarabandhu et.al shows 91.8 % of precision and 96.6% of recall whereas the technique given by j.yanget. al have lesser rate of about 84.90% and 90.0 % precision and recall. But our method shows 99.65% of preci-sion rate and 99.8% of recall. The other existing methods proposed by Kim et.al and J.Gllavata exhibits even smaller rate as given in the Table 1. The accuracy of the algorithm in the proposed method is computed by counting the number of correctly located characters, which is taken as the ground truth. The precision and recall rate are calculated by the equation (5) & (6) (5) (6) Comparisons with some existing methods are shown in table 1 which shows a clear improvement over existing methods. The performances of other methods shown are cited from the published works. The proposed system is implemented using Matlab[13]. Method Precision rate (%) Recall rate (%) Proposed method 99.65 99.8 Samarabandhu et. al[6] 91.8 96.6 J. yang et. al[11] 84.90 90.0 K.C.kim et al[10] 63.7 82.8 J.Gllavata et.al[3] 83.9 88.7 Table 1 Comparisons with existing methods Fig.6. The GUI showing the input image and the image with the text portion
  • 7. Automatic Text Extraction System For Complex Images DOI: 10.9790/0661-1803012329 www.iosrjournals.org 29 | Page V. Conclusion In this paper, a relatively simple, fast and effective method for text detection and extraction are proposed. The method uses DWT for the efficient working of the algorithm. This process requires less processing time which is mainly essential for real time applications and shows high precision rate. Most of methods fail when the characters are not aligned well or when the characters are too small. Those methods also result in some of missing characters when the characters have very poor contrast with respect to the background. But the pro-posed method is not sensitive to color or intensity of image, and also the uneven illumination and reflection effects. This can be used in large variety of application fields such as vehicle license plate detection, object identification, identification of various parts in industrial automation, mobile robot navigation which helps to detect text based land marks, analysis of technical papers with the help of maps, charts, and electric circuits etc. This algorithm is good at handling both scene text images and documents images effectively. Even though there are large numbers of algorithms in this area, it is observed that there is no single unified approach or algorithm that fits for all the applications. References [1] Xin Zhang, Fuchun Sun, Lei Gu “A Combined Al-gorithm for Video Text Extraction” Seventh Internation-al Conference on Fuzzy Systems and Knowledge Discov-ery,2010. [2] Ohya, A. Shio, and S. Akamatsu, “Recognizing Characters in Scene Images”, IEEE Transactions on Pat-tern Analysis and Machine Intelligence, 1994, 214-224. [3] Gllavata, R. Ewerth, and B. Freisleben, “A robust algorithm for text detection in images” Proceedings of the 3rd International Symposium on Image and Signal Pro-cessing and Analysis, pp.611– 616, ISPA, 2003. [4] Zhong, Yu.,Karu, K., and Jain, A.K.” Locating text in complex color images” Proceedings of the Third Inter-national Conference on Document Analysis and Recogni-tion1995, 1, 14-16: 146-149. [5] WonjunKim,Changick Kim, “A New Approach for Overlay Text Detection and Extraction From Complex Video Scene,” IEEE Transactions on Image Processing, V.18 , No.2, pp. 401 – 411, 2009. [6] Xiaoqing Liu, JagathSamarabandu “Multiscale Edge-Based Text Extraction from Complex Images,” International Conference on Multimedia and Expo, pp.1721-1724, 2006 [7] Xiao-Wei Zhang, Xiong-Bo Zheng, Zhi-Juan Weng, “Text Localization Algorithm Under Back-ground Image Using Wavelet Transforms”, Proceedings of the 2008 International Conference on Wavelet Analysis and Pattern Recognition, Hong Kong, pp.30- 31, Aug. 2008. [8] N. Otsu, “A Threshold Selection Method from Gray-Level Histograms,” IEEE Transactions on Systems, man and Cybernet, 1979. [9] Debapratim Sarkar, Raghunath Ghosh ,”A Bottom-Up Approach of Line Segmentation from Handwritten Text”2009. [10] K.C. Kim, H.R. Byun, Y.J. Song, Y.W. Choi, S.Y. Chi,K.K. Kim and Y.K Chung, Scene Text Extraction in Natural Scene Images using Hierarchical FeatureCom-bining and verification, Proceedings of the 17International Conference on Pattern Recognition (ICPR’04),IEEE. [11] J. Yang, J. Gao, Y. Zhang, X. Chen and A. Waibel, “AnAutomatic Sign Recognition and Translation Sys-tem”,Proceedings of the Workshop on Perceptive User Inter-faces(PUI'01), 2001, pp. 1-8. [12] A.K Jain,”Fundamentals of Digital Image Pro-cessing”,Englewood cliff, NJ: Prentice Hall, 1989.