SlideShare a Scribd company logo
1 of 5
Download to read offline
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 509
DESIGN AND DEVELOPMENT OF TESSERACT-OCR BASED ASSISTIVE
SYSTEM TO CONVERT CAPTURED TEXT INTO VOICE OUTPUT
G.ELUMALAI1,J.SUNDAR RAJAN2, P.SURYA PRAKASAH3, V.L.SUSRUTH4,P.K.SUDHARSANAN5
1 Associate Professor, Dept. of Electronics and Communication Engineering, Panimalar Engineering College,
Tamil Nadu, India
2,3,4,5 UG students, Dept. of Electronics and Communication Engineering, Panimalar Engineering College,
Tamil Nadu, India
---------------------------------------------------------------------***---------------------------------------------------------------------
Abstract - The biggest challenge faced by the blind people,
is their inability to view real life objects and to read. The only
efficient system that exists so far, is the braille system, that
enables the blind to read. This system is time consuming and
the time taken to recognize the text is long. Our aim here is to
reduce the time taken to read. In our work, using a Raspberry
Pi, we have designed a smart reader, so that the blind people
may read. The module that we have designed, either uses a
webcam or a mobile camera that is linked with a RaspberryPi,
to focus on a range of printed text. The OCR (OpticalCharacter
Recognition) package installed in raspberry pi tests it into a
digital article which is then subjected to skew modification,
segmentation, before feature extraction to perform sorting.
Our proposed project automatically focuses the regionsoftext
in the object, after which the text characters are thenlocalized
using a localization algorithm, that uses feature recognition
and edge pixel distribution using artificialneuralnetwork.The
text characters are then binarized into a machine readable
form by using an OCR algorithm called as Tesseract. The
recognized characters of the text are then converted into
audio format, so that they may be recognized by the blind
users.
Key Words: Embedded devices, Image processing, OCR,
Neural Networks, Raspberry Pi.
1. INTRODUCTION
The importance of reading in our society, today, cannot be
ignored. We could find printed text everywhere in the form
of medical prescription, newspapers, books, handouts,
medicine bottles and tablets and the like. While video
magnifiers, optical aids, and screen readers can help with
low vision and blind users to access documents, few devices
do exist that provide optimum access to daily-life hand held
objects such as fruit juice cartons and invention packages.
The most important causes for visual degradation include
diabetic retinopathy, aging factor, eye diseases such as
ocular tumors and accidents, which lead to an increaseinthe
number of visually disabled people, nowadays. Cataract is
leading cause of blindness and visual impairment. Mobile
applications are available these days, thataidinvisualization
and they form an integral part of the blind peoples lives.
Recent advances in mobile technology, digital camera,
computer vision and camera-based application make it
possible to support visually challenged persons by
developing camera-based application that combine
computer vision with other existing technology such as
optical character recognition (OCR) system. To detect text
information from image, practical difficulties exist, such as
non-uniform backgrounds, due to the large variations in
character font, size, texture, color, background, orientations,
and many other reasons. Text detection from scene/text
camera images is possible due to high resolution camera.
2. PROJECT OBJECTIVE
The biggest challenge faced by the blind peopletodayistheir
inability to read and recognize the real-world objects. The
only efficient method that exists so far is the braille method,
which enables them to read. It is somewhat time consuming
to read the text using a braille and also it might be not be
economically feasible. The objective of the project is to
create a portable, light-weight module that enablestheblind
to read. This module uses a webcam to capture the image of
the focused text, extract separate text characters from the
scene and then provide a speech output using the Tesseract
OCR.
3. LITERATURE SURVEY
In the field of Text To Speech conversion, researchisfocused
on the identification of the input text and the ability to
express it in a natural manner. Sentimental analysis
procedure(positive/neutral/negative) is used in order to
input features for expressive speech synthesis, where
identification of the input text is a front end task. Different
combinations of classifiers and textualfeaturesareevaluated
for the determination of the most appropriate adaptation
procedure. Semeval 2007 dataset and Twitter corpus are
used for the evaluation of the effectivenessof thisscheme of
Sentiment Analysis, which is appropriate for an expressive
Text To Speech scenario. Conducted experiments validate
the procedure proposed with respect to Sentiment Analysis
[1]. MD-TTS TC not only considers not only the contents of
the text but also its structure, in contrast to the topic of text
classification tasks. The proposal is validated by the
experiments that were conducted in terms of both
subjective(synthetic speech quality) and objective(TC
efficiency) evaluation criteria. [2] For the visually disabled
people, travel aids have been developed in the recent years
in the form of stereovision-capable devices. They are
however, bulky and assist the blind person in avoiding
obstacles only and they are head-mounted. We researched
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 510
about a wearable system that can not only assist in avoiding
obstacles, but at the same time supports live streaming of
videos over the 3G network. The module is comprised of an
eye-glass and an embedded device that is power efficient.
Two miniature camerasare used for imaging, on one end,by
the eye-glass. For the combination of stereo images, FPGA
and FIFO buffers are used. Also, live stereo matching on an
embedded processor is achieved by means of parallel
processing. Live streaming over 3G network, of the video
that is captured by the system is also supported. A person
with healthy eyesight can guide a visually impaired person,
with the help of this livevideo streaming [3]. We researched
about a system that helps the blind people to read from the
labelsand product packages of common handhelddevicesin
their daily lives, by means of a camera. The Region Of
Interest(ROI) in the camera view is obtained by using an
innovative motion-based approach, where the users were
asked to shake the object. This is done in order to isolate the
intended image from cluttered backgrounds and other
objects in the camera view. A mixture of Gaussian based,
background subtraction methodswere used by thismethod.
Text information is obtained from the extracted ROI, by a
combination of localization and recognition. Feature
extraction and the distribution of edge pixels are studied,
using Adaboost model, to automatically localize the text
characters in the ROI. Localized text characters are then
identified by using the OCR algorithms. Recognized text is
conveyed to the blind users using audio output. The
effectiveness of the system hardware is evaluated by using
the dataset that is obtained from ten blind people[4]. In TTS
technology, the ability is provided to a computer to speak. A
synthesizer is used to convert the text into speech, thus
making it eligible for further processing. By learningvarious
textbooks related to languages, language mastery was
obtained, traditionally. Due to the implied cost factor,
moving along this method was very difficult. Various
dictionaries were used by the existing system, to pronounce
the identified text with correct pronunciation . Only for a set
of words which are available in the dictionary, the operation
is supported. [5] Studies in blind people have shown the
activation of cortical areas that sub serve the vision, but it
has long been controversial whether blind people have
tactile acuity. We compared the passivetactileacuityofblind
and sighted subjects on a fully automatedgratingorientation
task and used multivariate Bayesian data analysis to
determine predictors of acuity. Irrespective oftheamountof
childhood vision, braille reading or light perception level,
acuity was superior in blind people. Acuity is better in
women than in men and it also decline with age. Acuity is
chiefly dependent on the force of contact between the
stimulus surface and the subject . The difference between
blind people and healthy sighted people varied a lot in spite
of the intragroup variation: the acuity of both sighted and
blind people of the same gender remained the same, except
that the acuity of a blind person was similar to that of a
sighted person who was 23 years younger. The results
suggest that cross modal plasticity may underlie tactile
acuity enhancement in blindness. [6]
4. EXISTING SYSTEM
A few systems already exist that help the blind people have
access to daily-life hand held objects, but they are largely
ineffective with respect to the focusarea of the labelling.For
example, portable bar-code readers do exist, but it is very
difficult for a blind person to locate a bar codeandthenpoint
the laser beam on that bar code. Also, to enable the blind
people to read, braille system exists. In spite of ignoring the
manufacturing cost, the system can be said to be largely
ineffective with regard to the processing speed, that is, the
speed at which the blind subject can identify theinformation
and then assimilate it is too low.
5. PROPOSED SYSTEM
The proposed system uses a Raspberry Pi board, an
ultrasonic sensor and a webcam to recognize the text in the
scene. In this, the webcam is focussed on the scene. A video
streaming is obtained, from which the images are captured
frame by frame. The images are refined in order to eliminate
any noise that is present in it. A feature called segmentation
is used in order to separate each character from other in the
text. Graphical details such as icons or logos, if any, are
eliminated. Each obtained character are compared with the
datasets that are created as a part of the Tesseract library.
The Tesseract OCR is the most efficient algorithm available
that checks for the obtained character in ten dimensions.
Once, the character is recognized, it must be made available
as an audio output. For this, we use a software calledfestival.
The festival is used to provide the audio output for the
recognized character. Apart from these features, an extra
feature is added, that enables the blind to know the type of
object that he/she interacts with.(a menu, newspaper and
the like). An ultrasonic sensor is included as a part of the
project, that makes the project obtain charactersonlywithin
a particular distance.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 511
Fig-1: Block diagram of the proposed system
6. SYSTEM DESIGN SPECIFICATION
6.1. Hardware specification
1. RASPBERRY PI:The RaspberryPi,usingtheARM1176JZF-S
core, is a credit card sized single computer on chip or SoC.
System on chip is a method by which all the electronic
components that are necessary to run a computer systemare
placed on a single chip. In order to start up, the Raspberry Pi
needsan OS.On board non-volatilememory,whichstoresthe
bootloaders are eliminated in the Raspberry Pi, with the aim
of cost reduction. Traditional embedded systems use Linux
kernels and file system. For this reason, an SD/MMCcardslot
is provided. Raspberry Pi will be executed according to the
application program and after the completion of successful
boot loading.
2. ULTRASONIC SENSORS: Ultrasonic sensors provide cost
effective, unique sensingmethods, which arenotprovidedby
any other technology, thus serving the market at large.
Application problems that are eithercostprohibitiveorthose
that couldn’t be solved using other sensors can be solved by
an ultrasonic sensor, by applying a wide variety of ultrasonic
transducers and a wide variety of frequency ranges.
Detection over distance are required by a variety of
applications in industrial sensing. The detection range of
ultrasonic sensors is quite high, as high as up to forty feet,
which limit switches and inductive switching cannot do.
Photo electric sensorscandetectlong distances,buttheylack
a wide area coverage or a wide area coverage which can be
obtained by using a numberof sensors. Migatron’sultrasonic
sensor cover both narrow and wide areas. Proper ultrasonic
transducer selection is what it all takes. Since there are wide
range of target materials available, only ultrasonic sensors
are independent of the target material composition . The
target material may be solid, liquid, porous, soft or woody in
nature and irrespective of its target composition, it can still
be detected. Distance measurement is easy and it can be
calculated by using the time from which theultrasonic signal
leavesthe signaltothe time theultrasonicreturnsbacktothe
sensor. This method is accurate to .05% of range which
equates to +or- .002 of an inch at a distance of 4 inches.
6.2. Software specifications:
1. OPEN CV: OpenCV (Open Source Computer VisionLibrary)
is a collectionof library functions, whose main objective is to
enable computer vision. The main objective of OpenCV is to
provide a framework forcomputer vision and to improvethe
readabilityofmachinesover commercialproducts. Sinceitis
BSD licensed, businesses can easily use and modify the code.
Over2500 sophisticated algorithmsarepresentinthelibrary,
which include both machine learning and computer vision
algorithms. Using the algorithms, the computer can identify
objects, identify faces, recognize moving objects, recognize
human actions, extract 3D models of objects, combine
different images to produce an entire image scene, red eye
removal, scenery recognition and augmented reality etc. can
be performed. Around 47 thousandmemberusercommunity
for OpenCV exists and OpenCV is widely used by many
commercial bodies, governmental institution and research
organizations.
2. NEURAL-NETWORKS: Neural networksare a collectionof
hardware and software modules that are patterned around
the neural networks of the human brain, in technical terms
of information technology. They are a variety of deep
learning technology that are also termed as artificial neural
networks. Complex signal processing and complex patter
recognition tasks are some of the commercialapplicationsof
the neural networks. Image processing and text recognition,
handwriting recognition, weather prediction and facial
recognition are some examples of the applications of neural
networks since 2000. A neural network usually involves a
large number of parallel operating processors arranged in
tiers. Raw input information are received in the first tier –
analogous to optic nerves in human visual processing.
Instead of receiving the output from the raw input, each tier
receives the output from the tier that precedes it. The last
tier produces the output of the system. Every node that
processes possess its own sphere of knowledge, including
what it was originally programmed with, or what is has
learnt or those set of rules which it has created for itself.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 512
3. OCR: Optical Character Recognition is the conversion of
either printed or handwritten characters into a machine
readable format(binary 0s and 1s). The text of concern may
be obtained either from an imageof a document,orascanned
copyof the documentor from areallife scene such asthetext
on signs and billboards(Fig.6 and Fig.7) or from subtitle text
underneath an image (such asthose thatoccur ina television
broadcast). It is used to enter the information from different
types of documents such as records, passports, cheques,
invoice, bills or any other such suitable type of
documentation. OCR is a widely used method for converting
the printed data into digital format, so that they may be
edited, searched , compressed, used in line display and in
machine related activitiessuch asdata mining, TTS, machine
learning and cognitive computing.
4. ESPEAK: A job is done by a computer in three distinct
stages, wherein the first stage is called input (where
information is often fed using a computer and a mouse),
processing (where the input isrespondedtobythecomputer,
say, by subtracting the numbers that you typed in or by
enhancing the colors on a photo you scanned), and output
(where you can see how the computer has processed the
input, wherein it is obtained on a computer screen or as a
printout on a paper).Espeak uses a technology called as
speech synthesis, wherein a text material is loudly read out
by the computer using a real or simulated voice via a
loudspeaker. This technologyis referredtoasTextToSpeech.
5. PYTHON: Python is an object oriented programming
language that is widely used in Pi development boards. It is
the preferred language for data analytics and data science. It
wasfound towards the end of the 1980sasa successor to the
ABClanguage byGuido Van Rossum inthe NationalResearch
Institute for Mathematicsand Computer Science.Python has
a clear syntax and it is developed with the intention of
making the code readable. All these haspaved the way for its
popularity since its inception. Pythonisahigh-levellanguage.
The code for python is largely written in simple English, that
makes it understandable to the user. The commands
provided to the Pi board are mostly in English.This isinstark
contrast to an assembler that more or less operates in a way
that is comfortable for the machine, thus making it
cumbersome for a normal human to read.Pythonisvaluedby
those whowant tolearnprogramming,on accountofitsclear
syntax and the fact that it is a high level programming
language. Raspberry Pi recommends Python for those who
are willing to learn embedded programming from scratch.
7. IMPLEMENTATION OF THE PROPOSED SYSTEM
Themainidea behinddeveloping the device is theportability
and ease that it could offer to the end user. Hence, it is
deliberately intended to be simple, so that it could be of light
weight and hence, easy to carry. The module uses a
Raspberry Pidevelopment board, whichis a creditcardsized
SoC and is quite efficient. A webcam is interfaced with this
board, in order to enable the transmission of the captured
stream of video to the Pi board. An ultrasonic sensor is used
as a range finder, that enables the module to detect text data
only within a specified range. The intended simplicity of the
module comes from the coding part, which is towards the
software side. The most efficient OCR algorithm that is
available is the Tesseract OCR,whose accuracyofdetectionis
pretty high. Since portability is intended and the Pi board
runs either using a PCpower supplyor an AC adapter,we are
using a power banktopoweronthe Pi board. Theworkingof
the device thusgoeson asfollows: The Raspberry Pi board is
powered ON, on startup, the code for identifying and
recognizingthe text gets executed(Fig.4). The camera is then
focused towards the intended text. Ultrasonic sensor is used
to enable thewebcam toread the textonlywithinastipulated
area. The Tesseract algorithm then identifies each character
of the text and feeds it to another software module called as
Espeak. Espeak converts thedetected data into audiooutput.
The audiooutput ismade available at the3mmearphonejack
in the Pi board. The arrangement of the device(Fig.2 and
Fig.3), the scanned text and the obtainedoutput are included
below(Fig.5):
Fig-2: Top view of the proposed system
Fig-3: Side view of the proposed system
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 513
Fig-4: Scanned text
Fig-5: Obtained Output(since the audio output cannot be
included, the output is included in visual form)
Fig-6: Scanned object
The object scanned by the web camera is orange which is
identified by the name of citrus using the Imagga database
(Fig.6).Thetesseract detects thescannedobjectanddetectsit
as citrus and displays the output(as shown in Fig.7).
Fig-7: Output
8. CONCLUSION AND FUTURE SCOPE
A text detection and recognition with speech output system
was successfully demonstrated on Android platform. This
system is very handy and useful for the visually challenged
persons. Comparedwith a PC platform,themobileplatformis
portable and more convenient to use. This system will be
helpful for visuallychallengedpersons to accessinformation
in written form and in the surrounding. It is useful to
understand the written text messages, warnings, and traffic
direction in voice form by converting it from Text to voice. It
is found that this system is capable of converting the sign
boards and other text into speech. There is further scope in
the development of this project. The project may further be
developed into/embedded along with the walking stick,
which blind peopleusuallyuse tonavigate.Apartfromthis,in
the future, smaller development boards may come into
existence or smaller camerasmay be developed,thusfurther
enhancing the portability of this device. Flexible PCB boards
is the latest trending technology in the field of printed circuit
boards. If we implement all these into the project, the size of
the project will be greatlyreduced, thusbeingahandysystem
for the blind people.
9.REFERENCES
[1]. Alexandre Trilla and Francesc Alías. (2013), “Sentence
Based Sentiment Analysis for Expressive Text-to-Speech”,
IEEE Transactions on Audio, Speech, and Language
Processing, Vol. 21, Issue. 2. pp. 223-233.
[2]. Alías F, Sevillano X, Socoró J. C, Gonzalvo X. (2008),
“Towards high-quality next-generation text-to-speech
synthesis”, IEEE Trans. Audio, Speech,LanguageProcess,Vol.
16, No. 7. pp. 1340-1354.
[3]. Balakrishnan G, Sainarayanan G, Nagarajan R. and
Yaacob S. (2007) “Wearable real-time stereo vision for the
visually impaired”, Vol. 14, No. 2, pp. 6–14.
[4]. Chucai Yi, Yingli Tian, Aries Arditi. (2014), “Portable
Camera-based Assistive Text and Product Label Reading
from Hand-held Objects for Blind Persons”,IEEE/ASME
Transactions on Mechatronics, Vol. 3, No. 2, pp. 1-10.
[5]. Deepa Jose V. and Sharan R. (2014), “A Novel Model for
Speech to Text Conversion”, International Refereed Journal
of Engineering and Science (IRJES) Vol.3, Issue.1, pp. 39-41.
[6]. Goldreich D. and Kanics I. M. (2003), “Tactile Acuity is
Enhanced in Blindness”, International Journal of Research
and Science, Vol. 23, No. 8, pp. 3439–3445.

More Related Content

What's hot

Eye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AIEye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AIDr. Amarjeet Singh
 
A review of consumer brain computer interface devices
A review of consumer brain computer interface devicesA review of consumer brain computer interface devices
A review of consumer brain computer interface devicesRex Yuan
 
IRJET- Wearable AI Device for Blind
IRJET- Wearable AI Device for BlindIRJET- Wearable AI Device for Blind
IRJET- Wearable AI Device for BlindIRJET Journal
 
Smart Tracking System
Smart Tracking SystemSmart Tracking System
Smart Tracking SystemIRJET Journal
 
IRJET- Deep Learning Techniques for Object Detection
IRJET-  	  Deep Learning Techniques for Object DetectionIRJET-  	  Deep Learning Techniques for Object Detection
IRJET- Deep Learning Techniques for Object DetectionIRJET Journal
 
Modelling Framework of a Neural Object Recognition
Modelling Framework of a Neural Object RecognitionModelling Framework of a Neural Object Recognition
Modelling Framework of a Neural Object RecognitionIJERA Editor
 
IRJET- Image to Text Conversion using Tesseract
IRJET-  	  Image to Text Conversion using TesseractIRJET-  	  Image to Text Conversion using Tesseract
IRJET- Image to Text Conversion using TesseractIRJET Journal
 
IRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually ImpairedIRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually ImpairedIRJET Journal
 
Resume_Saifullah Abro
Resume_Saifullah AbroResume_Saifullah Abro
Resume_Saifullah AbroSaif Abro
 
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...Kalle
 
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...ijtsrd
 
Application on character recognition system on road sign for visually impaire...
Application on character recognition system on road sign for visually impaire...Application on character recognition system on road sign for visually impaire...
Application on character recognition system on road sign for visually impaire...IJECEIAES
 
Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...IRJET Journal
 
Communication among blind, deaf and dumb People
Communication among blind, deaf and dumb PeopleCommunication among blind, deaf and dumb People
Communication among blind, deaf and dumb PeopleIJAEMSJORNAL
 
Multi-modal Asian Conversation Mobile Video Dataset for Recognition Task
Multi-modal Asian Conversation Mobile Video Dataset for Recognition TaskMulti-modal Asian Conversation Mobile Video Dataset for Recognition Task
Multi-modal Asian Conversation Mobile Video Dataset for Recognition TaskIJECEIAES
 
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATION
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATIONMULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATION
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATIONgerogepatton
 
Precision Physiotherapy & Sports Training: Part 1
Precision Physiotherapy & Sports Training: Part 1Precision Physiotherapy & Sports Training: Part 1
Precision Physiotherapy & Sports Training: Part 1PetteriTeikariPhD
 
IRJET- Facial Expression Recognition using Deep Learning: A Review
IRJET- Facial Expression Recognition using Deep Learning: A ReviewIRJET- Facial Expression Recognition using Deep Learning: A Review
IRJET- Facial Expression Recognition using Deep Learning: A ReviewIRJET Journal
 
Braille Technology
Braille TechnologyBraille Technology
Braille TechnologyKhuloodSaeed
 

What's hot (20)

Eye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AIEye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AI
 
A review of consumer brain computer interface devices
A review of consumer brain computer interface devicesA review of consumer brain computer interface devices
A review of consumer brain computer interface devices
 
IRJET- Wearable AI Device for Blind
IRJET- Wearable AI Device for BlindIRJET- Wearable AI Device for Blind
IRJET- Wearable AI Device for Blind
 
Smart Tracking System
Smart Tracking SystemSmart Tracking System
Smart Tracking System
 
IRJET- Deep Learning Techniques for Object Detection
IRJET-  	  Deep Learning Techniques for Object DetectionIRJET-  	  Deep Learning Techniques for Object Detection
IRJET- Deep Learning Techniques for Object Detection
 
Modelling Framework of a Neural Object Recognition
Modelling Framework of a Neural Object RecognitionModelling Framework of a Neural Object Recognition
Modelling Framework of a Neural Object Recognition
 
33 102-1-pb
33 102-1-pb33 102-1-pb
33 102-1-pb
 
IRJET- Image to Text Conversion using Tesseract
IRJET-  	  Image to Text Conversion using TesseractIRJET-  	  Image to Text Conversion using Tesseract
IRJET- Image to Text Conversion using Tesseract
 
IRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually ImpairedIRJET- Navigation and Camera Reading System for Visually Impaired
IRJET- Navigation and Camera Reading System for Visually Impaired
 
Resume_Saifullah Abro
Resume_Saifullah AbroResume_Saifullah Abro
Resume_Saifullah Abro
 
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...
Kinnunen Towards Task Independent Person Authentication Using Eye Movement Si...
 
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
 
Application on character recognition system on road sign for visually impaire...
Application on character recognition system on road sign for visually impaire...Application on character recognition system on road sign for visually impaire...
Application on character recognition system on road sign for visually impaire...
 
Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...Design of a Communication System using Sign Language aid for Differently Able...
Design of a Communication System using Sign Language aid for Differently Able...
 
Communication among blind, deaf and dumb People
Communication among blind, deaf and dumb PeopleCommunication among blind, deaf and dumb People
Communication among blind, deaf and dumb People
 
Multi-modal Asian Conversation Mobile Video Dataset for Recognition Task
Multi-modal Asian Conversation Mobile Video Dataset for Recognition TaskMulti-modal Asian Conversation Mobile Video Dataset for Recognition Task
Multi-modal Asian Conversation Mobile Video Dataset for Recognition Task
 
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATION
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATIONMULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATION
MULTI-LEVEL FEATURE FUSION BASED TRANSFER LEARNING FOR PERSON RE-IDENTIFICATION
 
Precision Physiotherapy & Sports Training: Part 1
Precision Physiotherapy & Sports Training: Part 1Precision Physiotherapy & Sports Training: Part 1
Precision Physiotherapy & Sports Training: Part 1
 
IRJET- Facial Expression Recognition using Deep Learning: A Review
IRJET- Facial Expression Recognition using Deep Learning: A ReviewIRJET- Facial Expression Recognition using Deep Learning: A Review
IRJET- Facial Expression Recognition using Deep Learning: A Review
 
Braille Technology
Braille TechnologyBraille Technology
Braille Technology
 

Similar to IRJET- Design and Development of Tesseract-OCR Based Assistive System to Convert Captured Text into Voice Output

An Assistive System for Visually Impaired People
An Assistive System for Visually Impaired PeopleAn Assistive System for Visually Impaired People
An Assistive System for Visually Impaired PeopleIRJET Journal
 
Visual Product Identification For Blind Peoples
Visual Product Identification For Blind PeoplesVisual Product Identification For Blind Peoples
Visual Product Identification For Blind PeoplesIRJET Journal
 
IRJETDevice for Location Finder and Text Reader for Visually Impaired People
IRJETDevice for Location Finder and Text Reader for Visually Impaired PeopleIRJETDevice for Location Finder and Text Reader for Visually Impaired People
IRJETDevice for Location Finder and Text Reader for Visually Impaired PeopleIRJET Journal
 
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry PiIRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry PiIRJET Journal
 
Product Label Reading System for visually challenged people
Product Label Reading System for visually challenged peopleProduct Label Reading System for visually challenged people
Product Label Reading System for visually challenged peopleIRJET Journal
 
OPTICAL CHARACTER RECOGNITION IN HEALTHCARE
OPTICAL CHARACTER RECOGNITION IN HEALTHCAREOPTICAL CHARACTER RECOGNITION IN HEALTHCARE
OPTICAL CHARACTER RECOGNITION IN HEALTHCAREIRJET Journal
 
Assistive Examination System for Visually Impaired
Assistive Examination System for Visually ImpairedAssistive Examination System for Visually Impaired
Assistive Examination System for Visually ImpairedEditor IJCATR
 
IRJET - A Review on Text Recognition for Visually Blind People
IRJET - A Review on Text Recognition for Visually Blind PeopleIRJET - A Review on Text Recognition for Visually Blind People
IRJET - A Review on Text Recognition for Visually Blind PeopleIRJET Journal
 
Review On Speech Recognition using Deep Learning
Review On Speech Recognition using Deep LearningReview On Speech Recognition using Deep Learning
Review On Speech Recognition using Deep LearningIRJET Journal
 
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...IRJET Journal
 
IRJET- Text Recognization of Product for Blind Person using MATLAB
IRJET- Text Recognization of Product for Blind Person using MATLABIRJET- Text Recognization of Product for Blind Person using MATLAB
IRJET- Text Recognization of Product for Blind Person using MATLABIRJET Journal
 
Drishyam - Virtual Eye for Blind
Drishyam - Virtual Eye for BlindDrishyam - Virtual Eye for Blind
Drishyam - Virtual Eye for BlindIRJET Journal
 
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...YogeshIJTSRD
 
IRJET - Smart Vision System for Visually Impaired People
IRJET -  	  Smart Vision System for Visually Impaired PeopleIRJET -  	  Smart Vision System for Visually Impaired People
IRJET - Smart Vision System for Visually Impaired PeopleIRJET Journal
 
Real-Time Sign Language Detector
Real-Time Sign Language DetectorReal-Time Sign Language Detector
Real-Time Sign Language DetectorIRJET Journal
 
GLOVE BASED GESTURE RECOGNITION USING IR SENSOR
GLOVE BASED GESTURE RECOGNITION USING IR SENSORGLOVE BASED GESTURE RECOGNITION USING IR SENSOR
GLOVE BASED GESTURE RECOGNITION USING IR SENSORIRJET Journal
 
DHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEDHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEIRJET Journal
 
DHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEDHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEIRJET Journal
 
A review of factors that impact the design of a glove based wearable devices
A review of factors that impact the design of a glove based wearable devicesA review of factors that impact the design of a glove based wearable devices
A review of factors that impact the design of a glove based wearable devicesIAESIJAI
 
IRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET Journal
 

Similar to IRJET- Design and Development of Tesseract-OCR Based Assistive System to Convert Captured Text into Voice Output (20)

An Assistive System for Visually Impaired People
An Assistive System for Visually Impaired PeopleAn Assistive System for Visually Impaired People
An Assistive System for Visually Impaired People
 
Visual Product Identification For Blind Peoples
Visual Product Identification For Blind PeoplesVisual Product Identification For Blind Peoples
Visual Product Identification For Blind Peoples
 
IRJETDevice for Location Finder and Text Reader for Visually Impaired People
IRJETDevice for Location Finder and Text Reader for Visually Impaired PeopleIRJETDevice for Location Finder and Text Reader for Visually Impaired People
IRJETDevice for Location Finder and Text Reader for Visually Impaired People
 
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry PiIRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
 
Product Label Reading System for visually challenged people
Product Label Reading System for visually challenged peopleProduct Label Reading System for visually challenged people
Product Label Reading System for visually challenged people
 
OPTICAL CHARACTER RECOGNITION IN HEALTHCARE
OPTICAL CHARACTER RECOGNITION IN HEALTHCAREOPTICAL CHARACTER RECOGNITION IN HEALTHCARE
OPTICAL CHARACTER RECOGNITION IN HEALTHCARE
 
Assistive Examination System for Visually Impaired
Assistive Examination System for Visually ImpairedAssistive Examination System for Visually Impaired
Assistive Examination System for Visually Impaired
 
IRJET - A Review on Text Recognition for Visually Blind People
IRJET - A Review on Text Recognition for Visually Blind PeopleIRJET - A Review on Text Recognition for Visually Blind People
IRJET - A Review on Text Recognition for Visually Blind People
 
Review On Speech Recognition using Deep Learning
Review On Speech Recognition using Deep LearningReview On Speech Recognition using Deep Learning
Review On Speech Recognition using Deep Learning
 
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
 
IRJET- Text Recognization of Product for Blind Person using MATLAB
IRJET- Text Recognization of Product for Blind Person using MATLABIRJET- Text Recognization of Product for Blind Person using MATLAB
IRJET- Text Recognization of Product for Blind Person using MATLAB
 
Drishyam - Virtual Eye for Blind
Drishyam - Virtual Eye for BlindDrishyam - Virtual Eye for Blind
Drishyam - Virtual Eye for Blind
 
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...
Recognizing of Text and Product Label from Hand Held Entity Intended for Visi...
 
IRJET - Smart Vision System for Visually Impaired People
IRJET -  	  Smart Vision System for Visually Impaired PeopleIRJET -  	  Smart Vision System for Visually Impaired People
IRJET - Smart Vision System for Visually Impaired People
 
Real-Time Sign Language Detector
Real-Time Sign Language DetectorReal-Time Sign Language Detector
Real-Time Sign Language Detector
 
GLOVE BASED GESTURE RECOGNITION USING IR SENSOR
GLOVE BASED GESTURE RECOGNITION USING IR SENSORGLOVE BASED GESTURE RECOGNITION USING IR SENSOR
GLOVE BASED GESTURE RECOGNITION USING IR SENSOR
 
DHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEDHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTE
 
DHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTEDHWANI- THE VOICE OF DEAF AND MUTE
DHWANI- THE VOICE OF DEAF AND MUTE
 
A review of factors that impact the design of a glove based wearable devices
A review of factors that impact the design of a glove based wearable devicesA review of factors that impact the design of a glove based wearable devices
A review of factors that impact the design of a glove based wearable devices
 
IRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN Methodology
 

More from IRJET Journal

TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...IRJET Journal
 
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURESTUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTUREIRJET Journal
 
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...IRJET Journal
 
Effect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsEffect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsIRJET Journal
 
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...IRJET Journal
 
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...IRJET Journal
 
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...IRJET Journal
 
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...IRJET Journal
 
A REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASA REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASIRJET Journal
 
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...IRJET Journal
 
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProP.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProIRJET Journal
 
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...IRJET Journal
 
Survey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemSurvey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemIRJET Journal
 
Review on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesReview on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesIRJET Journal
 
React based fullstack edtech web application
React based fullstack edtech web applicationReact based fullstack edtech web application
React based fullstack edtech web applicationIRJET Journal
 
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...IRJET Journal
 
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.IRJET Journal
 
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...IRJET Journal
 
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignMultistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignIRJET Journal
 
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...IRJET Journal
 

More from IRJET Journal (20)

TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
 
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURESTUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
 
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
 
Effect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsEffect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil Characteristics
 
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
 
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
 
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
 
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
 
A REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASA REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADAS
 
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
 
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProP.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
 
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
 
Survey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemSurvey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare System
 
Review on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesReview on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridges
 
React based fullstack edtech web application
React based fullstack edtech web applicationReact based fullstack edtech web application
React based fullstack edtech web application
 
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
 
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
 
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
 
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignMultistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
 
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
 

Recently uploaded

"Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments""Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments"mphochane1998
 
Path loss model, OKUMURA Model, Hata Model
Path loss model, OKUMURA Model, Hata ModelPath loss model, OKUMURA Model, Hata Model
Path loss model, OKUMURA Model, Hata ModelDrAjayKumarYadav4
 
Linux Systems Programming: Inter Process Communication (IPC) using Pipes
Linux Systems Programming: Inter Process Communication (IPC) using PipesLinux Systems Programming: Inter Process Communication (IPC) using Pipes
Linux Systems Programming: Inter Process Communication (IPC) using PipesRashidFaridChishti
 
8th International Conference on Soft Computing, Mathematics and Control (SMC ...
8th International Conference on Soft Computing, Mathematics and Control (SMC ...8th International Conference on Soft Computing, Mathematics and Control (SMC ...
8th International Conference on Soft Computing, Mathematics and Control (SMC ...josephjonse
 
Hospital management system project report.pdf
Hospital management system project report.pdfHospital management system project report.pdf
Hospital management system project report.pdfKamal Acharya
 
Max. shear stress theory-Maximum Shear Stress Theory ​ Maximum Distortional ...
Max. shear stress theory-Maximum Shear Stress Theory ​  Maximum Distortional ...Max. shear stress theory-Maximum Shear Stress Theory ​  Maximum Distortional ...
Max. shear stress theory-Maximum Shear Stress Theory ​ Maximum Distortional ...ronahami
 
Memory Interfacing of 8086 with DMA 8257
Memory Interfacing of 8086 with DMA 8257Memory Interfacing of 8086 with DMA 8257
Memory Interfacing of 8086 with DMA 8257subhasishdas79
 
Basic Electronics for diploma students as per technical education Kerala Syll...
Basic Electronics for diploma students as per technical education Kerala Syll...Basic Electronics for diploma students as per technical education Kerala Syll...
Basic Electronics for diploma students as per technical education Kerala Syll...ppkakm
 
Design For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startDesign For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startQuintin Balsdon
 
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXssuser89054b
 
Worksharing and 3D Modeling with Revit.pptx
Worksharing and 3D Modeling with Revit.pptxWorksharing and 3D Modeling with Revit.pptx
Worksharing and 3D Modeling with Revit.pptxMustafa Ahmed
 
Introduction to Data Visualization,Matplotlib.pdf
Introduction to Data Visualization,Matplotlib.pdfIntroduction to Data Visualization,Matplotlib.pdf
Introduction to Data Visualization,Matplotlib.pdfsumitt6_25730773
 
Introduction to Artificial Intelligence ( AI)
Introduction to Artificial Intelligence ( AI)Introduction to Artificial Intelligence ( AI)
Introduction to Artificial Intelligence ( AI)ChandrakantDivate1
 
UNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptxUNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptxkalpana413121
 
Kuwait City MTP kit ((+919101817206)) Buy Abortion Pills Kuwait
Kuwait City MTP kit ((+919101817206)) Buy Abortion Pills KuwaitKuwait City MTP kit ((+919101817206)) Buy Abortion Pills Kuwait
Kuwait City MTP kit ((+919101817206)) Buy Abortion Pills Kuwaitjaanualu31
 
👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...
👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...
👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...manju garg
 
Introduction to Geographic Information Systems
Introduction to Geographic Information SystemsIntroduction to Geographic Information Systems
Introduction to Geographic Information SystemsAnge Felix NSANZIYERA
 
Augmented Reality (AR) with Augin Software.pptx
Augmented Reality (AR) with Augin Software.pptxAugmented Reality (AR) with Augin Software.pptx
Augmented Reality (AR) with Augin Software.pptxMustafa Ahmed
 
Employee leave management system project.
Employee leave management system project.Employee leave management system project.
Employee leave management system project.Kamal Acharya
 
PE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and propertiesPE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and propertiessarkmank1
 

Recently uploaded (20)

"Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments""Lesotho Leaps Forward: A Chronicle of Transformative Developments"
"Lesotho Leaps Forward: A Chronicle of Transformative Developments"
 
Path loss model, OKUMURA Model, Hata Model
Path loss model, OKUMURA Model, Hata ModelPath loss model, OKUMURA Model, Hata Model
Path loss model, OKUMURA Model, Hata Model
 
Linux Systems Programming: Inter Process Communication (IPC) using Pipes
Linux Systems Programming: Inter Process Communication (IPC) using PipesLinux Systems Programming: Inter Process Communication (IPC) using Pipes
Linux Systems Programming: Inter Process Communication (IPC) using Pipes
 
8th International Conference on Soft Computing, Mathematics and Control (SMC ...
8th International Conference on Soft Computing, Mathematics and Control (SMC ...8th International Conference on Soft Computing, Mathematics and Control (SMC ...
8th International Conference on Soft Computing, Mathematics and Control (SMC ...
 
Hospital management system project report.pdf
Hospital management system project report.pdfHospital management system project report.pdf
Hospital management system project report.pdf
 
Max. shear stress theory-Maximum Shear Stress Theory ​ Maximum Distortional ...
Max. shear stress theory-Maximum Shear Stress Theory ​  Maximum Distortional ...Max. shear stress theory-Maximum Shear Stress Theory ​  Maximum Distortional ...
Max. shear stress theory-Maximum Shear Stress Theory ​ Maximum Distortional ...
 
Memory Interfacing of 8086 with DMA 8257
Memory Interfacing of 8086 with DMA 8257Memory Interfacing of 8086 with DMA 8257
Memory Interfacing of 8086 with DMA 8257
 
Basic Electronics for diploma students as per technical education Kerala Syll...
Basic Electronics for diploma students as per technical education Kerala Syll...Basic Electronics for diploma students as per technical education Kerala Syll...
Basic Electronics for diploma students as per technical education Kerala Syll...
 
Design For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startDesign For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the start
 
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
 
Worksharing and 3D Modeling with Revit.pptx
Worksharing and 3D Modeling with Revit.pptxWorksharing and 3D Modeling with Revit.pptx
Worksharing and 3D Modeling with Revit.pptx
 
Introduction to Data Visualization,Matplotlib.pdf
Introduction to Data Visualization,Matplotlib.pdfIntroduction to Data Visualization,Matplotlib.pdf
Introduction to Data Visualization,Matplotlib.pdf
 
Introduction to Artificial Intelligence ( AI)
Introduction to Artificial Intelligence ( AI)Introduction to Artificial Intelligence ( AI)
Introduction to Artificial Intelligence ( AI)
 
UNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptxUNIT 4 PTRP final Convergence in probability.pptx
UNIT 4 PTRP final Convergence in probability.pptx
 
Kuwait City MTP kit ((+919101817206)) Buy Abortion Pills Kuwait
Kuwait City MTP kit ((+919101817206)) Buy Abortion Pills KuwaitKuwait City MTP kit ((+919101817206)) Buy Abortion Pills Kuwait
Kuwait City MTP kit ((+919101817206)) Buy Abortion Pills Kuwait
 
👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...
👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...
👉 Yavatmal Call Girls Service Just Call 🍑👄6378878445 🍑👄 Top Class Call Girl S...
 
Introduction to Geographic Information Systems
Introduction to Geographic Information SystemsIntroduction to Geographic Information Systems
Introduction to Geographic Information Systems
 
Augmented Reality (AR) with Augin Software.pptx
Augmented Reality (AR) with Augin Software.pptxAugmented Reality (AR) with Augin Software.pptx
Augmented Reality (AR) with Augin Software.pptx
 
Employee leave management system project.
Employee leave management system project.Employee leave management system project.
Employee leave management system project.
 
PE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and propertiesPE 459 LECTURE 2- natural gas basic concepts and properties
PE 459 LECTURE 2- natural gas basic concepts and properties
 

IRJET- Design and Development of Tesseract-OCR Based Assistive System to Convert Captured Text into Voice Output

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 509 DESIGN AND DEVELOPMENT OF TESSERACT-OCR BASED ASSISTIVE SYSTEM TO CONVERT CAPTURED TEXT INTO VOICE OUTPUT G.ELUMALAI1,J.SUNDAR RAJAN2, P.SURYA PRAKASAH3, V.L.SUSRUTH4,P.K.SUDHARSANAN5 1 Associate Professor, Dept. of Electronics and Communication Engineering, Panimalar Engineering College, Tamil Nadu, India 2,3,4,5 UG students, Dept. of Electronics and Communication Engineering, Panimalar Engineering College, Tamil Nadu, India ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract - The biggest challenge faced by the blind people, is their inability to view real life objects and to read. The only efficient system that exists so far, is the braille system, that enables the blind to read. This system is time consuming and the time taken to recognize the text is long. Our aim here is to reduce the time taken to read. In our work, using a Raspberry Pi, we have designed a smart reader, so that the blind people may read. The module that we have designed, either uses a webcam or a mobile camera that is linked with a RaspberryPi, to focus on a range of printed text. The OCR (OpticalCharacter Recognition) package installed in raspberry pi tests it into a digital article which is then subjected to skew modification, segmentation, before feature extraction to perform sorting. Our proposed project automatically focuses the regionsoftext in the object, after which the text characters are thenlocalized using a localization algorithm, that uses feature recognition and edge pixel distribution using artificialneuralnetwork.The text characters are then binarized into a machine readable form by using an OCR algorithm called as Tesseract. The recognized characters of the text are then converted into audio format, so that they may be recognized by the blind users. Key Words: Embedded devices, Image processing, OCR, Neural Networks, Raspberry Pi. 1. INTRODUCTION The importance of reading in our society, today, cannot be ignored. We could find printed text everywhere in the form of medical prescription, newspapers, books, handouts, medicine bottles and tablets and the like. While video magnifiers, optical aids, and screen readers can help with low vision and blind users to access documents, few devices do exist that provide optimum access to daily-life hand held objects such as fruit juice cartons and invention packages. The most important causes for visual degradation include diabetic retinopathy, aging factor, eye diseases such as ocular tumors and accidents, which lead to an increaseinthe number of visually disabled people, nowadays. Cataract is leading cause of blindness and visual impairment. Mobile applications are available these days, thataidinvisualization and they form an integral part of the blind peoples lives. Recent advances in mobile technology, digital camera, computer vision and camera-based application make it possible to support visually challenged persons by developing camera-based application that combine computer vision with other existing technology such as optical character recognition (OCR) system. To detect text information from image, practical difficulties exist, such as non-uniform backgrounds, due to the large variations in character font, size, texture, color, background, orientations, and many other reasons. Text detection from scene/text camera images is possible due to high resolution camera. 2. PROJECT OBJECTIVE The biggest challenge faced by the blind peopletodayistheir inability to read and recognize the real-world objects. The only efficient method that exists so far is the braille method, which enables them to read. It is somewhat time consuming to read the text using a braille and also it might be not be economically feasible. The objective of the project is to create a portable, light-weight module that enablestheblind to read. This module uses a webcam to capture the image of the focused text, extract separate text characters from the scene and then provide a speech output using the Tesseract OCR. 3. LITERATURE SURVEY In the field of Text To Speech conversion, researchisfocused on the identification of the input text and the ability to express it in a natural manner. Sentimental analysis procedure(positive/neutral/negative) is used in order to input features for expressive speech synthesis, where identification of the input text is a front end task. Different combinations of classifiers and textualfeaturesareevaluated for the determination of the most appropriate adaptation procedure. Semeval 2007 dataset and Twitter corpus are used for the evaluation of the effectivenessof thisscheme of Sentiment Analysis, which is appropriate for an expressive Text To Speech scenario. Conducted experiments validate the procedure proposed with respect to Sentiment Analysis [1]. MD-TTS TC not only considers not only the contents of the text but also its structure, in contrast to the topic of text classification tasks. The proposal is validated by the experiments that were conducted in terms of both subjective(synthetic speech quality) and objective(TC efficiency) evaluation criteria. [2] For the visually disabled people, travel aids have been developed in the recent years in the form of stereovision-capable devices. They are however, bulky and assist the blind person in avoiding obstacles only and they are head-mounted. We researched
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 510 about a wearable system that can not only assist in avoiding obstacles, but at the same time supports live streaming of videos over the 3G network. The module is comprised of an eye-glass and an embedded device that is power efficient. Two miniature camerasare used for imaging, on one end,by the eye-glass. For the combination of stereo images, FPGA and FIFO buffers are used. Also, live stereo matching on an embedded processor is achieved by means of parallel processing. Live streaming over 3G network, of the video that is captured by the system is also supported. A person with healthy eyesight can guide a visually impaired person, with the help of this livevideo streaming [3]. We researched about a system that helps the blind people to read from the labelsand product packages of common handhelddevicesin their daily lives, by means of a camera. The Region Of Interest(ROI) in the camera view is obtained by using an innovative motion-based approach, where the users were asked to shake the object. This is done in order to isolate the intended image from cluttered backgrounds and other objects in the camera view. A mixture of Gaussian based, background subtraction methodswere used by thismethod. Text information is obtained from the extracted ROI, by a combination of localization and recognition. Feature extraction and the distribution of edge pixels are studied, using Adaboost model, to automatically localize the text characters in the ROI. Localized text characters are then identified by using the OCR algorithms. Recognized text is conveyed to the blind users using audio output. The effectiveness of the system hardware is evaluated by using the dataset that is obtained from ten blind people[4]. In TTS technology, the ability is provided to a computer to speak. A synthesizer is used to convert the text into speech, thus making it eligible for further processing. By learningvarious textbooks related to languages, language mastery was obtained, traditionally. Due to the implied cost factor, moving along this method was very difficult. Various dictionaries were used by the existing system, to pronounce the identified text with correct pronunciation . Only for a set of words which are available in the dictionary, the operation is supported. [5] Studies in blind people have shown the activation of cortical areas that sub serve the vision, but it has long been controversial whether blind people have tactile acuity. We compared the passivetactileacuityofblind and sighted subjects on a fully automatedgratingorientation task and used multivariate Bayesian data analysis to determine predictors of acuity. Irrespective oftheamountof childhood vision, braille reading or light perception level, acuity was superior in blind people. Acuity is better in women than in men and it also decline with age. Acuity is chiefly dependent on the force of contact between the stimulus surface and the subject . The difference between blind people and healthy sighted people varied a lot in spite of the intragroup variation: the acuity of both sighted and blind people of the same gender remained the same, except that the acuity of a blind person was similar to that of a sighted person who was 23 years younger. The results suggest that cross modal plasticity may underlie tactile acuity enhancement in blindness. [6] 4. EXISTING SYSTEM A few systems already exist that help the blind people have access to daily-life hand held objects, but they are largely ineffective with respect to the focusarea of the labelling.For example, portable bar-code readers do exist, but it is very difficult for a blind person to locate a bar codeandthenpoint the laser beam on that bar code. Also, to enable the blind people to read, braille system exists. In spite of ignoring the manufacturing cost, the system can be said to be largely ineffective with regard to the processing speed, that is, the speed at which the blind subject can identify theinformation and then assimilate it is too low. 5. PROPOSED SYSTEM The proposed system uses a Raspberry Pi board, an ultrasonic sensor and a webcam to recognize the text in the scene. In this, the webcam is focussed on the scene. A video streaming is obtained, from which the images are captured frame by frame. The images are refined in order to eliminate any noise that is present in it. A feature called segmentation is used in order to separate each character from other in the text. Graphical details such as icons or logos, if any, are eliminated. Each obtained character are compared with the datasets that are created as a part of the Tesseract library. The Tesseract OCR is the most efficient algorithm available that checks for the obtained character in ten dimensions. Once, the character is recognized, it must be made available as an audio output. For this, we use a software calledfestival. The festival is used to provide the audio output for the recognized character. Apart from these features, an extra feature is added, that enables the blind to know the type of object that he/she interacts with.(a menu, newspaper and the like). An ultrasonic sensor is included as a part of the project, that makes the project obtain charactersonlywithin a particular distance.
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 511 Fig-1: Block diagram of the proposed system 6. SYSTEM DESIGN SPECIFICATION 6.1. Hardware specification 1. RASPBERRY PI:The RaspberryPi,usingtheARM1176JZF-S core, is a credit card sized single computer on chip or SoC. System on chip is a method by which all the electronic components that are necessary to run a computer systemare placed on a single chip. In order to start up, the Raspberry Pi needsan OS.On board non-volatilememory,whichstoresthe bootloaders are eliminated in the Raspberry Pi, with the aim of cost reduction. Traditional embedded systems use Linux kernels and file system. For this reason, an SD/MMCcardslot is provided. Raspberry Pi will be executed according to the application program and after the completion of successful boot loading. 2. ULTRASONIC SENSORS: Ultrasonic sensors provide cost effective, unique sensingmethods, which arenotprovidedby any other technology, thus serving the market at large. Application problems that are eithercostprohibitiveorthose that couldn’t be solved using other sensors can be solved by an ultrasonic sensor, by applying a wide variety of ultrasonic transducers and a wide variety of frequency ranges. Detection over distance are required by a variety of applications in industrial sensing. The detection range of ultrasonic sensors is quite high, as high as up to forty feet, which limit switches and inductive switching cannot do. Photo electric sensorscandetectlong distances,buttheylack a wide area coverage or a wide area coverage which can be obtained by using a numberof sensors. Migatron’sultrasonic sensor cover both narrow and wide areas. Proper ultrasonic transducer selection is what it all takes. Since there are wide range of target materials available, only ultrasonic sensors are independent of the target material composition . The target material may be solid, liquid, porous, soft or woody in nature and irrespective of its target composition, it can still be detected. Distance measurement is easy and it can be calculated by using the time from which theultrasonic signal leavesthe signaltothe time theultrasonicreturnsbacktothe sensor. This method is accurate to .05% of range which equates to +or- .002 of an inch at a distance of 4 inches. 6.2. Software specifications: 1. OPEN CV: OpenCV (Open Source Computer VisionLibrary) is a collectionof library functions, whose main objective is to enable computer vision. The main objective of OpenCV is to provide a framework forcomputer vision and to improvethe readabilityofmachinesover commercialproducts. Sinceitis BSD licensed, businesses can easily use and modify the code. Over2500 sophisticated algorithmsarepresentinthelibrary, which include both machine learning and computer vision algorithms. Using the algorithms, the computer can identify objects, identify faces, recognize moving objects, recognize human actions, extract 3D models of objects, combine different images to produce an entire image scene, red eye removal, scenery recognition and augmented reality etc. can be performed. Around 47 thousandmemberusercommunity for OpenCV exists and OpenCV is widely used by many commercial bodies, governmental institution and research organizations. 2. NEURAL-NETWORKS: Neural networksare a collectionof hardware and software modules that are patterned around the neural networks of the human brain, in technical terms of information technology. They are a variety of deep learning technology that are also termed as artificial neural networks. Complex signal processing and complex patter recognition tasks are some of the commercialapplicationsof the neural networks. Image processing and text recognition, handwriting recognition, weather prediction and facial recognition are some examples of the applications of neural networks since 2000. A neural network usually involves a large number of parallel operating processors arranged in tiers. Raw input information are received in the first tier – analogous to optic nerves in human visual processing. Instead of receiving the output from the raw input, each tier receives the output from the tier that precedes it. The last tier produces the output of the system. Every node that processes possess its own sphere of knowledge, including what it was originally programmed with, or what is has learnt or those set of rules which it has created for itself.
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 512 3. OCR: Optical Character Recognition is the conversion of either printed or handwritten characters into a machine readable format(binary 0s and 1s). The text of concern may be obtained either from an imageof a document,orascanned copyof the documentor from areallife scene such asthetext on signs and billboards(Fig.6 and Fig.7) or from subtitle text underneath an image (such asthose thatoccur ina television broadcast). It is used to enter the information from different types of documents such as records, passports, cheques, invoice, bills or any other such suitable type of documentation. OCR is a widely used method for converting the printed data into digital format, so that they may be edited, searched , compressed, used in line display and in machine related activitiessuch asdata mining, TTS, machine learning and cognitive computing. 4. ESPEAK: A job is done by a computer in three distinct stages, wherein the first stage is called input (where information is often fed using a computer and a mouse), processing (where the input isrespondedtobythecomputer, say, by subtracting the numbers that you typed in or by enhancing the colors on a photo you scanned), and output (where you can see how the computer has processed the input, wherein it is obtained on a computer screen or as a printout on a paper).Espeak uses a technology called as speech synthesis, wherein a text material is loudly read out by the computer using a real or simulated voice via a loudspeaker. This technologyis referredtoasTextToSpeech. 5. PYTHON: Python is an object oriented programming language that is widely used in Pi development boards. It is the preferred language for data analytics and data science. It wasfound towards the end of the 1980sasa successor to the ABClanguage byGuido Van Rossum inthe NationalResearch Institute for Mathematicsand Computer Science.Python has a clear syntax and it is developed with the intention of making the code readable. All these haspaved the way for its popularity since its inception. Pythonisahigh-levellanguage. The code for python is largely written in simple English, that makes it understandable to the user. The commands provided to the Pi board are mostly in English.This isinstark contrast to an assembler that more or less operates in a way that is comfortable for the machine, thus making it cumbersome for a normal human to read.Pythonisvaluedby those whowant tolearnprogramming,on accountofitsclear syntax and the fact that it is a high level programming language. Raspberry Pi recommends Python for those who are willing to learn embedded programming from scratch. 7. IMPLEMENTATION OF THE PROPOSED SYSTEM Themainidea behinddeveloping the device is theportability and ease that it could offer to the end user. Hence, it is deliberately intended to be simple, so that it could be of light weight and hence, easy to carry. The module uses a Raspberry Pidevelopment board, whichis a creditcardsized SoC and is quite efficient. A webcam is interfaced with this board, in order to enable the transmission of the captured stream of video to the Pi board. An ultrasonic sensor is used as a range finder, that enables the module to detect text data only within a specified range. The intended simplicity of the module comes from the coding part, which is towards the software side. The most efficient OCR algorithm that is available is the Tesseract OCR,whose accuracyofdetectionis pretty high. Since portability is intended and the Pi board runs either using a PCpower supplyor an AC adapter,we are using a power banktopoweronthe Pi board. Theworkingof the device thusgoeson asfollows: The Raspberry Pi board is powered ON, on startup, the code for identifying and recognizingthe text gets executed(Fig.4). The camera is then focused towards the intended text. Ultrasonic sensor is used to enable thewebcam toread the textonlywithinastipulated area. The Tesseract algorithm then identifies each character of the text and feeds it to another software module called as Espeak. Espeak converts thedetected data into audiooutput. The audiooutput ismade available at the3mmearphonejack in the Pi board. The arrangement of the device(Fig.2 and Fig.3), the scanned text and the obtainedoutput are included below(Fig.5): Fig-2: Top view of the proposed system Fig-3: Side view of the proposed system
  • 5. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 04 | Apr-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 513 Fig-4: Scanned text Fig-5: Obtained Output(since the audio output cannot be included, the output is included in visual form) Fig-6: Scanned object The object scanned by the web camera is orange which is identified by the name of citrus using the Imagga database (Fig.6).Thetesseract detects thescannedobjectanddetectsit as citrus and displays the output(as shown in Fig.7). Fig-7: Output 8. CONCLUSION AND FUTURE SCOPE A text detection and recognition with speech output system was successfully demonstrated on Android platform. This system is very handy and useful for the visually challenged persons. Comparedwith a PC platform,themobileplatformis portable and more convenient to use. This system will be helpful for visuallychallengedpersons to accessinformation in written form and in the surrounding. It is useful to understand the written text messages, warnings, and traffic direction in voice form by converting it from Text to voice. It is found that this system is capable of converting the sign boards and other text into speech. There is further scope in the development of this project. The project may further be developed into/embedded along with the walking stick, which blind peopleusuallyuse tonavigate.Apartfromthis,in the future, smaller development boards may come into existence or smaller camerasmay be developed,thusfurther enhancing the portability of this device. Flexible PCB boards is the latest trending technology in the field of printed circuit boards. If we implement all these into the project, the size of the project will be greatlyreduced, thusbeingahandysystem for the blind people. 9.REFERENCES [1]. Alexandre Trilla and Francesc Alías. (2013), “Sentence Based Sentiment Analysis for Expressive Text-to-Speech”, IEEE Transactions on Audio, Speech, and Language Processing, Vol. 21, Issue. 2. pp. 223-233. [2]. Alías F, Sevillano X, Socoró J. C, Gonzalvo X. (2008), “Towards high-quality next-generation text-to-speech synthesis”, IEEE Trans. Audio, Speech,LanguageProcess,Vol. 16, No. 7. pp. 1340-1354. [3]. Balakrishnan G, Sainarayanan G, Nagarajan R. and Yaacob S. (2007) “Wearable real-time stereo vision for the visually impaired”, Vol. 14, No. 2, pp. 6–14. [4]. Chucai Yi, Yingli Tian, Aries Arditi. (2014), “Portable Camera-based Assistive Text and Product Label Reading from Hand-held Objects for Blind Persons”,IEEE/ASME Transactions on Mechatronics, Vol. 3, No. 2, pp. 1-10. [5]. Deepa Jose V. and Sharan R. (2014), “A Novel Model for Speech to Text Conversion”, International Refereed Journal of Engineering and Science (IRJES) Vol.3, Issue.1, pp. 39-41. [6]. Goldreich D. and Kanics I. M. (2003), “Tactile Acuity is Enhanced in Blindness”, International Journal of Research and Science, Vol. 23, No. 8, pp. 3439–3445.