SlideShare a Scribd company logo
1 of 9
Download to read offline
Journal of Science and Technology (JST)
Volume 2, Issue 4, April 2017, PP 33-41
www.jst.org.in ISSN: 2456 - 5660
www.jst.org.in 33 | Page
Smart Multilingual Sign Boards
Pravin A. Dhulekar1
, Niharika Prajapati2
, Tejal A. Tribhuvan3
, Karishma S.
Godse4
1
(Asst Professor, Department of Electronics and Telecommunication Engineering, Sandip Institute of
Technology and Research Center, Sandip Foundation, Nashik, India)
2.3,4
(Student, Department of Electronics and Telecommunication Engineering, Sandip Institute of Technology
and Research Center, Sandip Foundation, Nashik, India)
Abstract : The Project is based on design & implementation of smart hybrid system for street sign boards
recognition, text and speech conversions through character extraction and symbol matching. The default
language use to pronounce signs on the street boards is English. Here we are proposing a novel method to
convert identified character or symbol into multiple languages like Hindi, Marathi, Urdu, etc. This Project is
helpful to all starting from the visually impaired, the tourists, the illiterates and all the people who travel. The
system is accomplished with the speech pronunciation in different languages and to display on screen. This
Project has a multidisciplinary approach as it belongs to the domains like computer vision, speech processing,
& Google cloud platform. Computer vision is used for character and symbol extraction from sign boards.
Speech processing is used for text to speech conversion. GCP is used for multiple language conversion of
original extracted text. Further programming is done for real time pronunciation and displaying desired output.
Keywords - Street Sign Boards Recognition, Character Extraction, GCP, Symbol Matching, Computer Vision
I. INTRODUCTION
Communication is a very important tool in a human life. It is an essential requirement in this world to
survive. We can look back in the times of ignorance when no language was developed even than communication
existed in form of sign language and other forms. To overcome the difficulties faced in communication; this
paper has been implemented to achieve a real time system with feature/character/symbol extraction, text to
speech conversion and then text conversion into different languages. It starts with capturing any image may be
color, gray scale or monochrome via any video acquisition device and that image is stored in various file
formats (such as bmp, jpeg, tiff, etc.). After it is stored, image is read by the software and feature extraction is
performed. After character recognition is done, it is displayed and speech signal is generated which can be heard
via speakers. GCP here is used as real time language translation platform. This platform help us to get the
conversion of the extracted character in different languages , which are decided according to the users for e.g.
English, Hindi, Marathi, Urdu, etc., as we use the Google cloud platform there is none of the boundary for the
language which is displayed. This complete system is very helpful for visually impaired, illiterate, tourists and
everyone who travels.
To make the cities of India a smart city, we must focus on their needs and the greatest opportunities to
improve lives, increase security on roads, accident avoidance and safe driving.
II. LITERATURE REVIEW
A recent survey says that 285 million people are estimated to be visually impaired worldwide, 39
million are blind and 246 have low vision. Out of every 100, 82 are living with blindness are aged 50 and above
20% of all such impairment‟s still cannot be prevented or cured. In today‟s scenario, the overall literacy rate for
people aged 15 and above is 86.3% which means there are still 13.7% people illiterate around the world.
Starting from the most illiterate country like South Sudan, Afghanistan to the most literate countries like
Norway, Cuba, etc., everywhere there is a requirement of local language for the illiterate or the rural people to
communicate.
Over the past years, tourism has been growing rapidly strong and an underlying contributor to the
economic recovery. Around 1.1 billion people traveled abroad in 2014 and Europe is the most visited area in the
world. In 2015, 244 million people, or 3.3% of the world's population, lived outside their country‟s origin. It is
predicted that immigration rates will continue to increase over time and every time they travel, have to face with
different languages at different places.
Hence, for all such people across the globe a common solution, a speech generation system using
language translation software is implemented which will be useful to everyone. In [3] presented a work on
converting the multilingual text into speech for visually impaired. Basic working principle used here is u law
companding. This system used full for the blind and mute people. In [4] presented a paper based on the nature
Journal of Science and Technology
www.jst.org.in 34 | Page
prosody generation in the text to voice for English using the phonetic integration .This system is very
complicated to record and accumulate the word but is used the less memory space. In [5] worked on a system of
converting the text in the form of speech. This paper includes the recognized text and converted proper sound
signal. OCR techniques are implemented to get the output of the system. In [6] worked on the text to voice
conversion for the application on android environment. The basic objective of
this paper is extending the offline OCR technologies to the android platform. They proposed there system to
detect the text which modified to the multi-resolution and converted into different languages using language
translator. In [7] presented the basic idea of interfacing of speech to computer. The automatic speech recognition
(ASR) played the key role of taking the speech technology to the people. The system is classified further as
feature extraction & feature recognition. The aim was to study and to explore how neural network can be
implemented to recognize a particular word speech as an alternative to the traditional methods.
In [8] presented the voice recognition system automatically by machine means an easy mode of
communication between human & machine. The area of voice processing has implied to application like text to
speech processing, speech to text processing, telephonic communication call routing etc. Here various features
extraction, technique are discussed such as relative spectral processing (RASTA), linear predictive coding
(LPC) analysis and many more. In [9] presented the idea of an innovative and real time low cost technique that
makes user comfortable to hear the contents of the images or any kind of abstracted document instead of reading
through them. It has used a combination of OCR “Optical Character Recognition and TTS text to speech
synthesizer “.This system helps the visually impaired person to interact with the machine like the computer
easily through the vocal interface.
In [10] proposed a system which is mainly designed for the blind user , which have interest in the field
of education , this people study with the help of audio recording provided by NGO‟s or the Braille Books. This
will provide them to have an audio material of their own choice of any printed material of object. The whole
system mainly includes OCR “Optical Character Recognition “ which will perform/include the various
operations like thresholding, filtering, segmentation and many more.
Figure 1: Street Sign Boards
III. PROPOSED TECHNIQUE
To make a hybrid system, the methodology used here is Optical Character Recognition and language
translation. The images will be captured via web camera, then extracting text from image to convert that text
into speech. After that, there is further process on extracted text of language translation using Google translation
software on cloud. Then the whole system will be helpful to all starting from the visually impaired, the tourists,
the illiterates and all the people who travel.
IV. ALGORITHM
Text reading system has two main parts: image to text conversion and text to voice conversion. Image into text
and then that text into speech conversion is achieved using programming on MATLAB platform.
Figure 2: Flowchart
Journal of Science and Technology
www.jst.org.in 35 | Page
A. Work Flow
1) Image Acquisition Device: A computer vision system processes images acquired from an electronic
camera, which is like the human vision system where the brain processes images derived from the eyes[8].The
camera or web-cam is used as the acquisition device, which captures the image symbol or sign which will
ultimately help us to get the sample image for the further process and to get the symbol for recognition. Also we
get the video from this device through which we can extract the required area of the interest and can be
processed further for various operations to get the final results.
a) Web Camera:
Figure 3: HP Laptop Web Camera
Technical Specification:
 Screen Size: 15.6 inches
 Screen Resolution: 1366 x 768
 Frame rate up to 30 frames per second
2) Optical Character Recognition:This block is used for character extraction and symbol matching which is
included in algorithm and explained in detailed, this is the area where all the input images or videos captured by
camera or webcam are processed and then the feature extraction is done by various techniques. Here we also
perform the template matching ,if the character or the symbol is to be matched, then the operations are
performed as character/symbol to text conversion[5].
Syntax:
txt = ocr(I)
txt = ocr(I, roi)
[ __ ] = ocr( __ ,Name,Value)
Example:
businessCard = imread(„businessCard .png‟);
ocrResults = ocr(businessCard)
recognisedText = ocrResults.Text;
figure;
imshow(businessCard);
text(600, 150, recognizedText, „BackgroundColor‟, [1 1 1]);
Steps:
Step 1: Image acquisition: Different types of images like monochrome, gray scale and color, any image can
be used as input to any video acquisition like web camera whose primary operation is to sense and capture.
Step 2: Preprocessing: Not mandatory but preprocessing is done when the input image is not clear and
requires deblurring, denoising, resizing or reshaping.
Step 3: Detect MSER Regions: Maximally Stable Extremal Regions (MSER) is used as a method of blob
detection in images. Mathematically,
Journal of Science and Technology
www.jst.org.in 36 | Page
Let Q1,…,Qi-1,Qi,… be a series of nested extremal regions (Qi ϛ Qi+1). Extremal region Qi* is maximally
stable if and only if q(i)=|Qi+∆ Qi-∆|/|Qi| has a local minimum at i*. (Here | | denotes cardinality). ∆ ϵ S is a
parameter of the method.
The equation checks for regions that remain stable over a certain number of thresholds. If a region Qi+∆ is not
significantly larger than a region Qi-∆, region Qi is taken as a maximally stable region.
3) Text to Speech Conversion: Basically called as “speech synthesis”. In this block the output obtained from
OCR i.e. the matched image, sign or symbol is in the form of text which we have converted by symbol to text
conversion technique here it is converted into speech by using text to voice converter technique, this speech is
processed further for various
languages ,which will help the user to get the information about the total area which is of the required interest
4)Language Translation: For this, we are using Google Cloud Platform which allows us to use its application
program interface at a minimal price. We chose Google translation API to connect with the MATLAB by
creating our own console on GCP and generating an API key that we have used to send data to and get data
from the cloud itself.
5)Via-laptop: We have implement the complete system by using laptop. We use the in-build camera of the
laptop for capturing the images, and by doing preprocessing on the captured image, we got the required results
which are displayed on the screen and at the same time the result is spelled out by the speaker of laptop.
6)Speakers and Display Device: It is the laptop speakers and the screen which is used for getting the voice in the
different languages like Hindi, English, Marathi, Urdu etc. which will help the user to get the complete info in
his own language it will be able to pronounce the different language to make the user easy to get the basic info.
This info will help the visually impaired person which is ultimately the goal of the system. The display is used
for displaying the different types of language outputs, it gives better results and which is the basic need of the
system.
B. Optical Character Recognition
Figure 4: Algorithm of OCR
1) Input Image: Capturing the image may be on streets while travelling, in classroom while teaching, for
information reading while browsing is all done by any video acquisition device such as a web camera. Then the
image is read with the help of „imread‟ command. „imread‟ reads image from graphics file.
2) Pre-Processing: Pre-processing is not a mandatory step. It is used only when the image is corrupted. It
consists of a number of steps to make the raw data usable for recognizer. It is mainly used for removal of noise,
deblurring, resizing and reshaping of the image. MSER technique is also used for blob detection.
Journal of Science and Technology
www.jst.org.in 37 | Page
3) Feature Extraction: The objective of feature extraction is to capture the essential characteristics of the
symbols. The most precise way of obtaining a character is by its raster image. Other way is to uproot certain
features that still characterize symbols, but leaves out the unimportant attributes [5].
4) Template Matching/Symbol Matching: A simpler way of template matching uses a convolution mask
(also called as template) structured to a related feature of the search image, which we want to detect. This
technique can be easily performed on grey images or edge images. We will match the extracted images with the
collected data or the data base. Here the information is already stored with the use of programming and hence,
created the data base.
5) Character/Symbol Recognition: Character recognition deals with cropping of each line and letter,
performing correlation and writing to a file. Before performing correlation we have to load the already stored
model templates so that we can match the letters with the templates. After we got the character by character
segmentation we store the character image in a structure. For pattern recognition, SURF (Speeded Up Robust
Feature) technique is used to match scene image with the box image with a max ratio of 0.65 and a threshold of
minimum 25 interest points.
6) Character/Symbol to Speech Conversion : First, it transforms text including symbols like numbers or
abbreviations into its equivalent of written-out words. This process is often called text normalization, pre-
processing, or tokenization[9].
7) Display : Display are used to display the text when we capture image by the camera or any type of the
scanner processing it and convert that symbol and character into the text and display on the screen.
C. Pattern Recognition (Algorithm)
SURF feature extraction technique is used for pattern based recognition. In this template images are
already stored in the database called as the „box image‟ and the image which is being captured through camera
is known as the „scene image‟. First interest points are obtained from the scene image and then matching is done
with the box image with a minimum threshold of 25 points. Also the max ratio is set to 0.65 based to get the
most accurate results. Also the matched result is shown by the matching lines clearly in the figure placed
below. The red and the green color points show the interest points. Algorithm of SURF technique is
as follows:
1. Covert images to gray-scale.
2. Extract SURF interest points.
3. Obtain features.
4. Match features.
5. Get 50 strongest features.
6. Match the number of strongest features with each image.
7. Take the ratio of- number of features matched/ number of strongest features (which is 25)
Figure 5: Matching of Box & Scene image
Journal of Science and Technology
www.jst.org.in 38 | Page
V. MATHEMATICAL ANALYSIS
1. Standard Data Set
This are the standard data base images which shows the100% accuracy and the following table shows
the response time for the image to give the output, which is very less.
Table 1: Standard Based
Sr.
No.
Name on Sign Board Accuracy Response
Time
1. Horns Prohibited 100% 0.217834
seconds
2. No Entry 100% 0.220310
seconds
3. No Parking 100% 0.198523
seconds
4. Overtaking
Prohibited
100% 0.234994
seconds
2.Pattern Based Data Set
The pattern based recognition shows the following Accuracy and the response time. Hence it shows the good
accuracy of the pattern or symbol based sign boards on which the processing is done.
Table 2: Pattern Based
Sr.
No.
Name on Sign
Board
Accuracy Response
Time
1. Men At Work 99% 0.298642
seconds
2. School Ahead Go
Slow
97% 0.383578
seconds
3. No U Turn 96% 0.379566
seconds
4. Turn Left 98% 0.286479
seconds
3.Real Time Data Set
This is the results of the real time data images which are captured by the user and shows 93% to 95%
accuracy. Also the response time is mentioned.
Table 3: Real Time Based
Sr.
No.
Name on Sign Board Accuracy Response
Time
1. U Turn Prohibited 93% 0.400865
seconds
2. Straight Prohibited 95% 0.389267
seconds
Journal of Science and Technology
www.jst.org.in 39 | Page
3. Do Not Honk Your
Horn
91% 0.305811
seconds
4. Overtaking
Prohibited
95% 0.477447
seconds
VI. RESULTS AND DISSCUSSIONS
Algorithm is piloted and tested and hence the results got generated as shown below:
This is the GUI of the complete project run on MATLAB which includes standard sign board
recognition based on both character and pattern recognition. Also real time sign board recognition is performed
Figure 6: Result Image I
Case 1(a): Result shows that each and every character from the image is extracted successfully irrespective
of their fonts using OCR & also voice is generated and displayed in three different languages..
Figure 7: Result Image II Case I(a)
Case 1(b): Result shows that how pattern of the captured image is matched with the template image & also
voice is generated and displayed in three different languages.
Journal of Science and Technology
www.jst.org.in 40 | Page
Figure 8: Result Image III (Case I(b)
Case 2: Result shows that each and every character from the image is extracted successfully in real time
using OCR & also voice is generated in three different languages and displayed for the same on real time
platform.
Figure 9: Result Image IV (Case II)
VII. CONCLUSION
This paper is an effort to implement an innovative robust approach for character extraction and text to
voice conversion of different images using optical character recognition and text to speech synthesis technology.
A user friendly, cost effective, reliable to all and applicable in the real time system is achieved. Using this
methodology, we can read text from a document, street sign boards, web page or also an e-Book and can
generate synthesized speech through any system i.e. computer's speakers or phone‟s speaker. In developed
software, the use of computer vision has set all policies of the characters corresponding to each and every
alphabet extraction, matching, its pronunciation, and the way it is used in grammar and dictionary. Speech
processing has given robust output for various images. Other application of this system includes such as making
information browsing for people who do not have the ability to read or write. This approach can be used in part
as well. If requirement is only for text conversion then it is possible or else text to speech conversion is also
done easily. People with vision impairment or visual dyslexia or complete blindness can use this approach for
reading the documents, books and also while travelling. People with vocal impairment or complete dumb person
can utilize this approach to turn typed words into vocalization. Tourists having language barrier can also use this
approach for understanding different languages. People travelling in cars and buses can save their time using
this feature. Experiments have been performed to test the text and speech generation system and good results
have been achieved. The work is done very effectively for the symbol extraction. We got required results as
provided in the above concept, the symbol extraction is done by using the SURF algorithm. Hence got the
suitable outputs and can overcome the problem of such a sign boards which are only in symbol format. We got
the outputs of the all the 3 methods as for the standard data base 100%, for the pattern base we acquired 96% to
98% and for the real time data set 91% to 95% accuracy with less response time . The developed system is now
able to deal with any sign boards of the RTO which is the basic need of the society, In this way we also
contribute to the concept of SMART CITY by contributing towards the SMART TRANSPORTATION.
Journal of Science and Technology
www.jst.org.in 41 | Page
REFERENCES
[1] Preeti.Kale, S. T. Gandhe, G. M. Phade, Pravin A. Dhulekar, “Enhancement of old images and documents by Digital Image Processing
Techniques” Presented in IEEE International Conference on Communication, Information & Computing Technology (ICCICT),
Mumbai, India. 16-17 January 2015.
[2] Harshada H. Chitte, Pravin A. Dhulekar, “Human Action Recognition for Video Surveillance” Presented in International Conference
on engineering confluence, Equinox, Mumbai, Oct 2014.
[3] Suraj Mallik, Rajesh Mehra,”Speech To Text Conversion For Visually Impaired Person Using Μ Law Companding ” Publish In Iosr
Journal Of Electronics And Communication Engineering (Iosr-Jece) E-Issn: 2278-2834,P- Issn: 2278-8735.Volume 10, Issue 6, Ver. Ii
(Nov - Dec .2015).
[4] Mrs. S. D. Suryawanshi, Mrs. R. R. Itkarkar, Mr. D. T. Mane ,“ High Quality Text To Speech Synthesizer Using Phonetic Integration
”Publish In International Journal Of Advanced Research In Electronics And Communication Engineering (Ijarece) Volume 3, Issue 2,
February 2014.
[5] Pooja Chandran, Aravind S, Jisha Gopinath And Saranya S S ,” Design And Implementation Of Speech Generation System Using
Matlab ” Publish In International Journal Of Engineering And Innovative Technology (Ijeit) Volume 4, Issue 6, December 2014.
[6] Devika Sharma ,Ranju Kanwar ,” Text To Speech Conversion With Language Translator Under Android Environment ” Publish In
International Journal Of Emerging Research In Management &Technology Issn: 2278-9359(Volume-4,Issue-6).
[7] Siddhant C. Joshi, Dr. A.N.Cheeran “Matlab Based Back-Propagation Neural Network For Automatic Speech Recognition”, IJAREEIE(An
Iso 3297: 2007 Certified Organization)Vol. 3, Issue 7, July 2014.
[8] Siddhant C. Joshi, Dr. A.N.Cheeran, “A Comparative Study of Feature Extraction Techniques for Speech Recognition System” ISSN:
2319-8753, IJIRSET ,Vol. 3, Issue 12, December 2014.
[9] K.Kalaivani, R.Praveena,V.Anjalipriya,R.Srimeena, “Real Time Implementation Of Image Recognition And Text To Speech
Conversion”,IJAERT,Volume 2 Issue 6, September 2014, Issn No.: 2348 – 8190.
[10] Kalyani Mangale, Hemangi Mhaske, Priyanka Wankhade, Vivek Niwane, “Printed Text To Audio Converter Using OCR”, International
Journal Of Computer Applications (0975 – 8887) (NCETACt-2015).
[11] Pravin A.Dhuleka,Niharika Prajapati, Tejal A Tribhuvan, Karishma S. Godse.” Automatic Voice Generation System After Street Board
Identification For Visually Impaired” ICSPICC-2016,Bambori,Jalgoan,
Maharashtra.22-24 December 2016.
[12] Mark Nixon, Alberto Aguado, Feature Extraction and Image Processing, 2nd
Edition. Sunil Jadhav, P. A. Dhulekar, Dr.G.M.Phade,
“Hybrid Optical Character Recognition Method for Recognition of Text in Images”, IEEE FIFTH International Conference on
Computing of Power ,Energy & Communication” ICCPEIC -2016, April 20th and 21th – 2016

More Related Content

What's hot

Braille Technology
Braille TechnologyBraille Technology
Braille TechnologyKhuloodSaeed
 
Implementation of humanoid robot with using the concept of synthetic brain
Implementation of humanoid robot with using the concept of synthetic brainImplementation of humanoid robot with using the concept of synthetic brain
Implementation of humanoid robot with using the concept of synthetic braineSAT Journals
 
Implementation of humanoid robot with using the
Implementation of humanoid robot with using theImplementation of humanoid robot with using the
Implementation of humanoid robot with using theeSAT Publishing House
 
Projects1920 c6 fina;
Projects1920 c6 fina;Projects1920 c6 fina;
Projects1920 c6 fina;TabassumBanu5
 
Speech Based Search Engine System Control and User Interaction
Speech Based Search Engine System Control and User  InteractionSpeech Based Search Engine System Control and User  Interaction
Speech Based Search Engine System Control and User InteractionIOSR Journals
 
A Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign LanguageA Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign Languageijsrd.com
 
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET Journal
 
Speech Recognition: Transcription and transformation of human speech
Speech Recognition: Transcription and transformation of human speechSpeech Recognition: Transcription and transformation of human speech
Speech Recognition: Transcription and transformation of human speechSubmissionResearchpa
 
IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...
IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...
IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...IRJET Journal
 
The role of speech technology in biometrics, forensics and man-machine interface
The role of speech technology in biometrics, forensics and man-machine interfaceThe role of speech technology in biometrics, forensics and man-machine interface
The role of speech technology in biometrics, forensics and man-machine interfaceIJECEIAES
 
IRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET Journal
 
Optical character recognition for Ge'ez characters
Optical character recognition for Ge'ez charactersOptical character recognition for Ge'ez characters
Optical character recognition for Ge'ez charactershadmac
 
Paper Design and Simulation of Communication Aid For Disabled based on region...
Paper Design and Simulation of Communication Aid For Disabled based on region...Paper Design and Simulation of Communication Aid For Disabled based on region...
Paper Design and Simulation of Communication Aid For Disabled based on region...Syambabu Choda
 
An embedded module as “virtual tongue”
An embedded module as “virtual tongue”An embedded module as “virtual tongue”
An embedded module as “virtual tongue”ijistjournal
 
Speech Recognition
Speech RecognitionSpeech Recognition
Speech Recognitionfathitarek
 
Handwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text ConversionHandwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text Conversionijtsrd
 
Voice input and speech recognition system in tourism/social media
Voice input and speech recognition system in tourism/social mediaVoice input and speech recognition system in tourism/social media
Voice input and speech recognition system in tourism/social mediacidroypaes
 

What's hot (20)

Braille Technology
Braille TechnologyBraille Technology
Braille Technology
 
Implementation of humanoid robot with using the concept of synthetic brain
Implementation of humanoid robot with using the concept of synthetic brainImplementation of humanoid robot with using the concept of synthetic brain
Implementation of humanoid robot with using the concept of synthetic brain
 
Implementation of humanoid robot with using the
Implementation of humanoid robot with using theImplementation of humanoid robot with using the
Implementation of humanoid robot with using the
 
Projects1920 c6 fina;
Projects1920 c6 fina;Projects1920 c6 fina;
Projects1920 c6 fina;
 
Speech Based Search Engine System Control and User Interaction
Speech Based Search Engine System Control and User  InteractionSpeech Based Search Engine System Control and User  Interaction
Speech Based Search Engine System Control and User Interaction
 
30
3030
30
 
A Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign LanguageA Translation Device for the Vision Based Sign Language
A Translation Device for the Vision Based Sign Language
 
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
 
Speech Recognition: Transcription and transformation of human speech
Speech Recognition: Transcription and transformation of human speechSpeech Recognition: Transcription and transformation of human speech
Speech Recognition: Transcription and transformation of human speech
 
IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...
IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...
IRJET- Communication System for Blind, Deaf and Dumb People using Internet of...
 
The role of speech technology in biometrics, forensics and man-machine interface
The role of speech technology in biometrics, forensics and man-machine interfaceThe role of speech technology in biometrics, forensics and man-machine interface
The role of speech technology in biometrics, forensics and man-machine interface
 
IRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN Methodology
 
Optical character recognition for Ge'ez characters
Optical character recognition for Ge'ez charactersOptical character recognition for Ge'ez characters
Optical character recognition for Ge'ez characters
 
Paper Design and Simulation of Communication Aid For Disabled based on region...
Paper Design and Simulation of Communication Aid For Disabled based on region...Paper Design and Simulation of Communication Aid For Disabled based on region...
Paper Design and Simulation of Communication Aid For Disabled based on region...
 
An embedded module as “virtual tongue”
An embedded module as “virtual tongue”An embedded module as “virtual tongue”
An embedded module as “virtual tongue”
 
CRC Final Report
CRC Final ReportCRC Final Report
CRC Final Report
 
Ocr 1
Ocr 1Ocr 1
Ocr 1
 
Speech Recognition
Speech RecognitionSpeech Recognition
Speech Recognition
 
Handwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text ConversionHandwritten Text Recognition and Digital Text Conversion
Handwritten Text Recognition and Digital Text Conversion
 
Voice input and speech recognition system in tourism/social media
Voice input and speech recognition system in tourism/social mediaVoice input and speech recognition system in tourism/social media
Voice input and speech recognition system in tourism/social media
 

Similar to 5.smart multilingual sign boards

VOICE BASED E-MAIL
VOICE BASED E-MAILVOICE BASED E-MAIL
VOICE BASED E-MAILStudentRocks
 
Text Detection and Recognition with Speech Output for Visually Challenged Per...
Text Detection and Recognition with Speech Output for Visually Challenged Per...Text Detection and Recognition with Speech Output for Visually Challenged Per...
Text Detection and Recognition with Speech Output for Visually Challenged Per...IJERA Editor
 
IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...
IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...
IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...IRJET Journal
 
IRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for BlindsIRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for BlindsIRJET Journal
 
Assistive Examination System for Visually Impaired
Assistive Examination System for Visually ImpairedAssistive Examination System for Visually Impaired
Assistive Examination System for Visually ImpairedEditor IJCATR
 
Pakistan sign language to Urdu translator using Kinect
Pakistan sign language to Urdu translator using KinectPakistan sign language to Urdu translator using Kinect
Pakistan sign language to Urdu translator using KinectCSITiaesprime
 
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry PiIRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry PiIRJET Journal
 
A survey on Enhancements in Speech Recognition
A survey on Enhancements in Speech RecognitionA survey on Enhancements in Speech Recognition
A survey on Enhancements in Speech RecognitionIRJET Journal
 
Procedia Computer Science 94 ( 2016 ) 295 – 301 Avail.docx
 Procedia Computer Science   94  ( 2016 )  295 – 301 Avail.docx Procedia Computer Science   94  ( 2016 )  295 – 301 Avail.docx
Procedia Computer Science 94 ( 2016 ) 295 – 301 Avail.docxaryan532920
 
IRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras ModelIRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras ModelIRJET Journal
 
Artificial Intelligence for Speech Recognition
Artificial Intelligence for Speech RecognitionArtificial Intelligence for Speech Recognition
Artificial Intelligence for Speech RecognitionRHIMRJ Journal
 
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...IRJET Journal
 
IRJET - Sign Language Converter
IRJET -  	  Sign Language ConverterIRJET -  	  Sign Language Converter
IRJET - Sign Language ConverterIRJET Journal
 
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...IRJET Journal
 
IRJET - Sign Language Text to Speech Converter using Image Processing and...
IRJET -  	  Sign Language Text to Speech Converter using Image Processing and...IRJET -  	  Sign Language Text to Speech Converter using Image Processing and...
IRJET - Sign Language Text to Speech Converter using Image Processing and...IRJET Journal
 
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...IRJET Journal
 
DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...
DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...
DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...IJITCA Journal
 

Similar to 5.smart multilingual sign boards (20)

VOICE BASED E-MAIL
VOICE BASED E-MAILVOICE BASED E-MAIL
VOICE BASED E-MAIL
 
Text Detection and Recognition with Speech Output for Visually Challenged Per...
Text Detection and Recognition with Speech Output for Visually Challenged Per...Text Detection and Recognition with Speech Output for Visually Challenged Per...
Text Detection and Recognition with Speech Output for Visually Challenged Per...
 
IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...
IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...
IRJET - Language Linguist using Image Processing on Intelligent Transport Sys...
 
IRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for BlindsIRJET- Virtual Vision for Blinds
IRJET- Virtual Vision for Blinds
 
Assistive Examination System for Visually Impaired
Assistive Examination System for Visually ImpairedAssistive Examination System for Visually Impaired
Assistive Examination System for Visually Impaired
 
Pakistan sign language to Urdu translator using Kinect
Pakistan sign language to Urdu translator using KinectPakistan sign language to Urdu translator using Kinect
Pakistan sign language to Urdu translator using Kinect
 
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry PiIRJET- Text Reading for Visually Impaired Person using Raspberry Pi
IRJET- Text Reading for Visually Impaired Person using Raspberry Pi
 
A survey on Enhancements in Speech Recognition
A survey on Enhancements in Speech RecognitionA survey on Enhancements in Speech Recognition
A survey on Enhancements in Speech Recognition
 
Procedia Computer Science 94 ( 2016 ) 295 – 301 Avail.docx
 Procedia Computer Science   94  ( 2016 )  295 – 301 Avail.docx Procedia Computer Science   94  ( 2016 )  295 – 301 Avail.docx
Procedia Computer Science 94 ( 2016 ) 295 – 301 Avail.docx
 
IRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras ModelIRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras Model
 
Artificial Intelligence for Speech Recognition
Artificial Intelligence for Speech RecognitionArtificial Intelligence for Speech Recognition
Artificial Intelligence for Speech Recognition
 
A Robust Embedded Based String Recognition for Visually Impaired People
A Robust Embedded Based String Recognition for Visually Impaired PeopleA Robust Embedded Based String Recognition for Visually Impaired People
A Robust Embedded Based String Recognition for Visually Impaired People
 
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
LIP READING - AN EFFICIENT CROSS AUDIO-VIDEO RECOGNITION USING 3D CONVOLUTION...
 
IRJET - Sign Language Converter
IRJET -  	  Sign Language ConverterIRJET -  	  Sign Language Converter
IRJET - Sign Language Converter
 
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...
INDIAN SIGN LANGUAGE TRANSLATION FOR HARD-OF-HEARING AND HARD-OF-SPEAKING COM...
 
voice browser
voice browservoice browser
voice browser
 
IRJET - Sign Language Text to Speech Converter using Image Processing and...
IRJET -  	  Sign Language Text to Speech Converter using Image Processing and...IRJET -  	  Sign Language Text to Speech Converter using Image Processing and...
IRJET - Sign Language Text to Speech Converter using Image Processing and...
 
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
 
OCR.pptx
OCR.pptxOCR.pptx
OCR.pptx
 
DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...
DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...
DEVELOPMENT OF AN INTEGRATED TOOL THAT SUMMARRIZE AND PRODUCE THE SIGN LANGUA...
 

Recently uploaded

chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learningmisbanausheenparvam
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ
 
microprocessor 8085 and its interfacing
microprocessor 8085  and its interfacingmicroprocessor 8085  and its interfacing
microprocessor 8085 and its interfacingjaychoudhary37
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVRajaP95
 
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfCCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfAsst.prof M.Gokilavani
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024hassan khalil
 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionDr.Costas Sachpazis
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfAsst.prof M.Gokilavani
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024Mark Billinghurst
 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSKurinjimalarL3
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidNikhilNagaraju
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxwendy cai
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoão Esperancinha
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Dr.Costas Sachpazis
 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxPoojaBan
 
Introduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxIntroduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxvipinkmenon1
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort servicejennyeacort
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 

Recently uploaded (20)

chaitra-1.pptx fake news detection using machine learning
chaitra-1.pptx  fake news detection using machine learningchaitra-1.pptx  fake news detection using machine learning
chaitra-1.pptx fake news detection using machine learning
 
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
VICTOR MAESTRE RAMIREZ - Planetary Defender on NASA's Double Asteroid Redirec...
 
microprocessor 8085 and its interfacing
microprocessor 8085  and its interfacingmicroprocessor 8085  and its interfacing
microprocessor 8085 and its interfacing
 
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptxExploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
 
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdfCCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
CCS355 Neural Network & Deep Learning UNIT III notes and Question bank .pdf
 
Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024Architect Hassan Khalil Portfolio for 2024
Architect Hassan Khalil Portfolio for 2024
 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024
 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
 
main PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfidmain PPT.pptx of girls hostel security using rfid
main PPT.pptx of girls hostel security using rfid
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptx
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
 
Heart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptxHeart Disease Prediction using machine learning.pptx
Heart Disease Prediction using machine learning.pptx
 
Introduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxIntroduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptx
 
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
 
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANVI) Koregaon Park Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
 

5.smart multilingual sign boards

  • 1. Journal of Science and Technology (JST) Volume 2, Issue 4, April 2017, PP 33-41 www.jst.org.in ISSN: 2456 - 5660 www.jst.org.in 33 | Page Smart Multilingual Sign Boards Pravin A. Dhulekar1 , Niharika Prajapati2 , Tejal A. Tribhuvan3 , Karishma S. Godse4 1 (Asst Professor, Department of Electronics and Telecommunication Engineering, Sandip Institute of Technology and Research Center, Sandip Foundation, Nashik, India) 2.3,4 (Student, Department of Electronics and Telecommunication Engineering, Sandip Institute of Technology and Research Center, Sandip Foundation, Nashik, India) Abstract : The Project is based on design & implementation of smart hybrid system for street sign boards recognition, text and speech conversions through character extraction and symbol matching. The default language use to pronounce signs on the street boards is English. Here we are proposing a novel method to convert identified character or symbol into multiple languages like Hindi, Marathi, Urdu, etc. This Project is helpful to all starting from the visually impaired, the tourists, the illiterates and all the people who travel. The system is accomplished with the speech pronunciation in different languages and to display on screen. This Project has a multidisciplinary approach as it belongs to the domains like computer vision, speech processing, & Google cloud platform. Computer vision is used for character and symbol extraction from sign boards. Speech processing is used for text to speech conversion. GCP is used for multiple language conversion of original extracted text. Further programming is done for real time pronunciation and displaying desired output. Keywords - Street Sign Boards Recognition, Character Extraction, GCP, Symbol Matching, Computer Vision I. INTRODUCTION Communication is a very important tool in a human life. It is an essential requirement in this world to survive. We can look back in the times of ignorance when no language was developed even than communication existed in form of sign language and other forms. To overcome the difficulties faced in communication; this paper has been implemented to achieve a real time system with feature/character/symbol extraction, text to speech conversion and then text conversion into different languages. It starts with capturing any image may be color, gray scale or monochrome via any video acquisition device and that image is stored in various file formats (such as bmp, jpeg, tiff, etc.). After it is stored, image is read by the software and feature extraction is performed. After character recognition is done, it is displayed and speech signal is generated which can be heard via speakers. GCP here is used as real time language translation platform. This platform help us to get the conversion of the extracted character in different languages , which are decided according to the users for e.g. English, Hindi, Marathi, Urdu, etc., as we use the Google cloud platform there is none of the boundary for the language which is displayed. This complete system is very helpful for visually impaired, illiterate, tourists and everyone who travels. To make the cities of India a smart city, we must focus on their needs and the greatest opportunities to improve lives, increase security on roads, accident avoidance and safe driving. II. LITERATURE REVIEW A recent survey says that 285 million people are estimated to be visually impaired worldwide, 39 million are blind and 246 have low vision. Out of every 100, 82 are living with blindness are aged 50 and above 20% of all such impairment‟s still cannot be prevented or cured. In today‟s scenario, the overall literacy rate for people aged 15 and above is 86.3% which means there are still 13.7% people illiterate around the world. Starting from the most illiterate country like South Sudan, Afghanistan to the most literate countries like Norway, Cuba, etc., everywhere there is a requirement of local language for the illiterate or the rural people to communicate. Over the past years, tourism has been growing rapidly strong and an underlying contributor to the economic recovery. Around 1.1 billion people traveled abroad in 2014 and Europe is the most visited area in the world. In 2015, 244 million people, or 3.3% of the world's population, lived outside their country‟s origin. It is predicted that immigration rates will continue to increase over time and every time they travel, have to face with different languages at different places. Hence, for all such people across the globe a common solution, a speech generation system using language translation software is implemented which will be useful to everyone. In [3] presented a work on converting the multilingual text into speech for visually impaired. Basic working principle used here is u law companding. This system used full for the blind and mute people. In [4] presented a paper based on the nature
  • 2. Journal of Science and Technology www.jst.org.in 34 | Page prosody generation in the text to voice for English using the phonetic integration .This system is very complicated to record and accumulate the word but is used the less memory space. In [5] worked on a system of converting the text in the form of speech. This paper includes the recognized text and converted proper sound signal. OCR techniques are implemented to get the output of the system. In [6] worked on the text to voice conversion for the application on android environment. The basic objective of this paper is extending the offline OCR technologies to the android platform. They proposed there system to detect the text which modified to the multi-resolution and converted into different languages using language translator. In [7] presented the basic idea of interfacing of speech to computer. The automatic speech recognition (ASR) played the key role of taking the speech technology to the people. The system is classified further as feature extraction & feature recognition. The aim was to study and to explore how neural network can be implemented to recognize a particular word speech as an alternative to the traditional methods. In [8] presented the voice recognition system automatically by machine means an easy mode of communication between human & machine. The area of voice processing has implied to application like text to speech processing, speech to text processing, telephonic communication call routing etc. Here various features extraction, technique are discussed such as relative spectral processing (RASTA), linear predictive coding (LPC) analysis and many more. In [9] presented the idea of an innovative and real time low cost technique that makes user comfortable to hear the contents of the images or any kind of abstracted document instead of reading through them. It has used a combination of OCR “Optical Character Recognition and TTS text to speech synthesizer “.This system helps the visually impaired person to interact with the machine like the computer easily through the vocal interface. In [10] proposed a system which is mainly designed for the blind user , which have interest in the field of education , this people study with the help of audio recording provided by NGO‟s or the Braille Books. This will provide them to have an audio material of their own choice of any printed material of object. The whole system mainly includes OCR “Optical Character Recognition “ which will perform/include the various operations like thresholding, filtering, segmentation and many more. Figure 1: Street Sign Boards III. PROPOSED TECHNIQUE To make a hybrid system, the methodology used here is Optical Character Recognition and language translation. The images will be captured via web camera, then extracting text from image to convert that text into speech. After that, there is further process on extracted text of language translation using Google translation software on cloud. Then the whole system will be helpful to all starting from the visually impaired, the tourists, the illiterates and all the people who travel. IV. ALGORITHM Text reading system has two main parts: image to text conversion and text to voice conversion. Image into text and then that text into speech conversion is achieved using programming on MATLAB platform. Figure 2: Flowchart
  • 3. Journal of Science and Technology www.jst.org.in 35 | Page A. Work Flow 1) Image Acquisition Device: A computer vision system processes images acquired from an electronic camera, which is like the human vision system where the brain processes images derived from the eyes[8].The camera or web-cam is used as the acquisition device, which captures the image symbol or sign which will ultimately help us to get the sample image for the further process and to get the symbol for recognition. Also we get the video from this device through which we can extract the required area of the interest and can be processed further for various operations to get the final results. a) Web Camera: Figure 3: HP Laptop Web Camera Technical Specification:  Screen Size: 15.6 inches  Screen Resolution: 1366 x 768  Frame rate up to 30 frames per second 2) Optical Character Recognition:This block is used for character extraction and symbol matching which is included in algorithm and explained in detailed, this is the area where all the input images or videos captured by camera or webcam are processed and then the feature extraction is done by various techniques. Here we also perform the template matching ,if the character or the symbol is to be matched, then the operations are performed as character/symbol to text conversion[5]. Syntax: txt = ocr(I) txt = ocr(I, roi) [ __ ] = ocr( __ ,Name,Value) Example: businessCard = imread(„businessCard .png‟); ocrResults = ocr(businessCard) recognisedText = ocrResults.Text; figure; imshow(businessCard); text(600, 150, recognizedText, „BackgroundColor‟, [1 1 1]); Steps: Step 1: Image acquisition: Different types of images like monochrome, gray scale and color, any image can be used as input to any video acquisition like web camera whose primary operation is to sense and capture. Step 2: Preprocessing: Not mandatory but preprocessing is done when the input image is not clear and requires deblurring, denoising, resizing or reshaping. Step 3: Detect MSER Regions: Maximally Stable Extremal Regions (MSER) is used as a method of blob detection in images. Mathematically,
  • 4. Journal of Science and Technology www.jst.org.in 36 | Page Let Q1,…,Qi-1,Qi,… be a series of nested extremal regions (Qi ϛ Qi+1). Extremal region Qi* is maximally stable if and only if q(i)=|Qi+∆ Qi-∆|/|Qi| has a local minimum at i*. (Here | | denotes cardinality). ∆ ϵ S is a parameter of the method. The equation checks for regions that remain stable over a certain number of thresholds. If a region Qi+∆ is not significantly larger than a region Qi-∆, region Qi is taken as a maximally stable region. 3) Text to Speech Conversion: Basically called as “speech synthesis”. In this block the output obtained from OCR i.e. the matched image, sign or symbol is in the form of text which we have converted by symbol to text conversion technique here it is converted into speech by using text to voice converter technique, this speech is processed further for various languages ,which will help the user to get the information about the total area which is of the required interest 4)Language Translation: For this, we are using Google Cloud Platform which allows us to use its application program interface at a minimal price. We chose Google translation API to connect with the MATLAB by creating our own console on GCP and generating an API key that we have used to send data to and get data from the cloud itself. 5)Via-laptop: We have implement the complete system by using laptop. We use the in-build camera of the laptop for capturing the images, and by doing preprocessing on the captured image, we got the required results which are displayed on the screen and at the same time the result is spelled out by the speaker of laptop. 6)Speakers and Display Device: It is the laptop speakers and the screen which is used for getting the voice in the different languages like Hindi, English, Marathi, Urdu etc. which will help the user to get the complete info in his own language it will be able to pronounce the different language to make the user easy to get the basic info. This info will help the visually impaired person which is ultimately the goal of the system. The display is used for displaying the different types of language outputs, it gives better results and which is the basic need of the system. B. Optical Character Recognition Figure 4: Algorithm of OCR 1) Input Image: Capturing the image may be on streets while travelling, in classroom while teaching, for information reading while browsing is all done by any video acquisition device such as a web camera. Then the image is read with the help of „imread‟ command. „imread‟ reads image from graphics file. 2) Pre-Processing: Pre-processing is not a mandatory step. It is used only when the image is corrupted. It consists of a number of steps to make the raw data usable for recognizer. It is mainly used for removal of noise, deblurring, resizing and reshaping of the image. MSER technique is also used for blob detection.
  • 5. Journal of Science and Technology www.jst.org.in 37 | Page 3) Feature Extraction: The objective of feature extraction is to capture the essential characteristics of the symbols. The most precise way of obtaining a character is by its raster image. Other way is to uproot certain features that still characterize symbols, but leaves out the unimportant attributes [5]. 4) Template Matching/Symbol Matching: A simpler way of template matching uses a convolution mask (also called as template) structured to a related feature of the search image, which we want to detect. This technique can be easily performed on grey images or edge images. We will match the extracted images with the collected data or the data base. Here the information is already stored with the use of programming and hence, created the data base. 5) Character/Symbol Recognition: Character recognition deals with cropping of each line and letter, performing correlation and writing to a file. Before performing correlation we have to load the already stored model templates so that we can match the letters with the templates. After we got the character by character segmentation we store the character image in a structure. For pattern recognition, SURF (Speeded Up Robust Feature) technique is used to match scene image with the box image with a max ratio of 0.65 and a threshold of minimum 25 interest points. 6) Character/Symbol to Speech Conversion : First, it transforms text including symbols like numbers or abbreviations into its equivalent of written-out words. This process is often called text normalization, pre- processing, or tokenization[9]. 7) Display : Display are used to display the text when we capture image by the camera or any type of the scanner processing it and convert that symbol and character into the text and display on the screen. C. Pattern Recognition (Algorithm) SURF feature extraction technique is used for pattern based recognition. In this template images are already stored in the database called as the „box image‟ and the image which is being captured through camera is known as the „scene image‟. First interest points are obtained from the scene image and then matching is done with the box image with a minimum threshold of 25 points. Also the max ratio is set to 0.65 based to get the most accurate results. Also the matched result is shown by the matching lines clearly in the figure placed below. The red and the green color points show the interest points. Algorithm of SURF technique is as follows: 1. Covert images to gray-scale. 2. Extract SURF interest points. 3. Obtain features. 4. Match features. 5. Get 50 strongest features. 6. Match the number of strongest features with each image. 7. Take the ratio of- number of features matched/ number of strongest features (which is 25) Figure 5: Matching of Box & Scene image
  • 6. Journal of Science and Technology www.jst.org.in 38 | Page V. MATHEMATICAL ANALYSIS 1. Standard Data Set This are the standard data base images which shows the100% accuracy and the following table shows the response time for the image to give the output, which is very less. Table 1: Standard Based Sr. No. Name on Sign Board Accuracy Response Time 1. Horns Prohibited 100% 0.217834 seconds 2. No Entry 100% 0.220310 seconds 3. No Parking 100% 0.198523 seconds 4. Overtaking Prohibited 100% 0.234994 seconds 2.Pattern Based Data Set The pattern based recognition shows the following Accuracy and the response time. Hence it shows the good accuracy of the pattern or symbol based sign boards on which the processing is done. Table 2: Pattern Based Sr. No. Name on Sign Board Accuracy Response Time 1. Men At Work 99% 0.298642 seconds 2. School Ahead Go Slow 97% 0.383578 seconds 3. No U Turn 96% 0.379566 seconds 4. Turn Left 98% 0.286479 seconds 3.Real Time Data Set This is the results of the real time data images which are captured by the user and shows 93% to 95% accuracy. Also the response time is mentioned. Table 3: Real Time Based Sr. No. Name on Sign Board Accuracy Response Time 1. U Turn Prohibited 93% 0.400865 seconds 2. Straight Prohibited 95% 0.389267 seconds
  • 7. Journal of Science and Technology www.jst.org.in 39 | Page 3. Do Not Honk Your Horn 91% 0.305811 seconds 4. Overtaking Prohibited 95% 0.477447 seconds VI. RESULTS AND DISSCUSSIONS Algorithm is piloted and tested and hence the results got generated as shown below: This is the GUI of the complete project run on MATLAB which includes standard sign board recognition based on both character and pattern recognition. Also real time sign board recognition is performed Figure 6: Result Image I Case 1(a): Result shows that each and every character from the image is extracted successfully irrespective of their fonts using OCR & also voice is generated and displayed in three different languages.. Figure 7: Result Image II Case I(a) Case 1(b): Result shows that how pattern of the captured image is matched with the template image & also voice is generated and displayed in three different languages.
  • 8. Journal of Science and Technology www.jst.org.in 40 | Page Figure 8: Result Image III (Case I(b) Case 2: Result shows that each and every character from the image is extracted successfully in real time using OCR & also voice is generated in three different languages and displayed for the same on real time platform. Figure 9: Result Image IV (Case II) VII. CONCLUSION This paper is an effort to implement an innovative robust approach for character extraction and text to voice conversion of different images using optical character recognition and text to speech synthesis technology. A user friendly, cost effective, reliable to all and applicable in the real time system is achieved. Using this methodology, we can read text from a document, street sign boards, web page or also an e-Book and can generate synthesized speech through any system i.e. computer's speakers or phone‟s speaker. In developed software, the use of computer vision has set all policies of the characters corresponding to each and every alphabet extraction, matching, its pronunciation, and the way it is used in grammar and dictionary. Speech processing has given robust output for various images. Other application of this system includes such as making information browsing for people who do not have the ability to read or write. This approach can be used in part as well. If requirement is only for text conversion then it is possible or else text to speech conversion is also done easily. People with vision impairment or visual dyslexia or complete blindness can use this approach for reading the documents, books and also while travelling. People with vocal impairment or complete dumb person can utilize this approach to turn typed words into vocalization. Tourists having language barrier can also use this approach for understanding different languages. People travelling in cars and buses can save their time using this feature. Experiments have been performed to test the text and speech generation system and good results have been achieved. The work is done very effectively for the symbol extraction. We got required results as provided in the above concept, the symbol extraction is done by using the SURF algorithm. Hence got the suitable outputs and can overcome the problem of such a sign boards which are only in symbol format. We got the outputs of the all the 3 methods as for the standard data base 100%, for the pattern base we acquired 96% to 98% and for the real time data set 91% to 95% accuracy with less response time . The developed system is now able to deal with any sign boards of the RTO which is the basic need of the society, In this way we also contribute to the concept of SMART CITY by contributing towards the SMART TRANSPORTATION.
  • 9. Journal of Science and Technology www.jst.org.in 41 | Page REFERENCES [1] Preeti.Kale, S. T. Gandhe, G. M. Phade, Pravin A. Dhulekar, “Enhancement of old images and documents by Digital Image Processing Techniques” Presented in IEEE International Conference on Communication, Information & Computing Technology (ICCICT), Mumbai, India. 16-17 January 2015. [2] Harshada H. Chitte, Pravin A. Dhulekar, “Human Action Recognition for Video Surveillance” Presented in International Conference on engineering confluence, Equinox, Mumbai, Oct 2014. [3] Suraj Mallik, Rajesh Mehra,”Speech To Text Conversion For Visually Impaired Person Using Μ Law Companding ” Publish In Iosr Journal Of Electronics And Communication Engineering (Iosr-Jece) E-Issn: 2278-2834,P- Issn: 2278-8735.Volume 10, Issue 6, Ver. Ii (Nov - Dec .2015). [4] Mrs. S. D. Suryawanshi, Mrs. R. R. Itkarkar, Mr. D. T. Mane ,“ High Quality Text To Speech Synthesizer Using Phonetic Integration ”Publish In International Journal Of Advanced Research In Electronics And Communication Engineering (Ijarece) Volume 3, Issue 2, February 2014. [5] Pooja Chandran, Aravind S, Jisha Gopinath And Saranya S S ,” Design And Implementation Of Speech Generation System Using Matlab ” Publish In International Journal Of Engineering And Innovative Technology (Ijeit) Volume 4, Issue 6, December 2014. [6] Devika Sharma ,Ranju Kanwar ,” Text To Speech Conversion With Language Translator Under Android Environment ” Publish In International Journal Of Emerging Research In Management &Technology Issn: 2278-9359(Volume-4,Issue-6). [7] Siddhant C. Joshi, Dr. A.N.Cheeran “Matlab Based Back-Propagation Neural Network For Automatic Speech Recognition”, IJAREEIE(An Iso 3297: 2007 Certified Organization)Vol. 3, Issue 7, July 2014. [8] Siddhant C. Joshi, Dr. A.N.Cheeran, “A Comparative Study of Feature Extraction Techniques for Speech Recognition System” ISSN: 2319-8753, IJIRSET ,Vol. 3, Issue 12, December 2014. [9] K.Kalaivani, R.Praveena,V.Anjalipriya,R.Srimeena, “Real Time Implementation Of Image Recognition And Text To Speech Conversion”,IJAERT,Volume 2 Issue 6, September 2014, Issn No.: 2348 – 8190. [10] Kalyani Mangale, Hemangi Mhaske, Priyanka Wankhade, Vivek Niwane, “Printed Text To Audio Converter Using OCR”, International Journal Of Computer Applications (0975 – 8887) (NCETACt-2015). [11] Pravin A.Dhuleka,Niharika Prajapati, Tejal A Tribhuvan, Karishma S. Godse.” Automatic Voice Generation System After Street Board Identification For Visually Impaired” ICSPICC-2016,Bambori,Jalgoan, Maharashtra.22-24 December 2016. [12] Mark Nixon, Alberto Aguado, Feature Extraction and Image Processing, 2nd Edition. Sunil Jadhav, P. A. Dhulekar, Dr.G.M.Phade, “Hybrid Optical Character Recognition Method for Recognition of Text in Images”, IEEE FIFTH International Conference on Computing of Power ,Energy & Communication” ICCPEIC -2016, April 20th and 21th – 2016