SlideShare a Scribd company logo
1 of 7
Download to read offline
International Journal of Trend in Scientific Research and Development (IJTSRD)
Volume 4 Issue 4, June 2020 Available Online: www.ijtsrd.com e-ISSN: 2456 – 6470
@ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1641
Godeye: An Efficient System for Blinds
Sabin Khader1, Meerakrishna M R1, Reshma Roy1, Willson Joseph C2
1UG Scholar, 2Assistant Professor,
1,2Department of Computer Science, Sahrdaya College of Engineering and Technology, Thrissur, Kerala, India
ABSTRACT
The technology is growing vastly. Everyone in humanity hassomelimitations.
One of those limitations is visual disability. So we are here with a system that
helps the visually disabled people. The framework here contains object
detection with voice assistance within an app and a hardwarepartattachedto
the blinds stick for distance calculation. The app is designed to support the
blind person to explore freely anywhere he wants. The working of the
framework begins by surveilling the situations around the user and
distinguishing them utilizing a camera. The app will then detect the objects
present in the input video frame by using the SSD algorithm comparingitwith
the trained model. The video captured is partitioned into grids to detect the
object/obstacle. In this way, the subtleties of the object detected can be
achieved and along with it distance measurement can also be calculatedusing
specific algorithms. A Text to Speech (TTS) converter is utilized for changing
over the data about the object detected into an audio speech format. The
framework/application passes on the scene which the blind peopleisgoingin
his/her territorial language with the snap of a catch. The technologiesutilized
here makes the framework execution effective.
KEYWORDS: visual impairment, android phone, object/obstacle detection, text-
to-speech
How to cite this paper: Sabin Khader |
Meerakrishna M R | Reshma Roy | Willson
Joseph C "Godeye: An Efficient System for
Blinds" Published in
International Journal
of Trend in Scientific
Research and
Development(ijtsrd),
ISSN: 2456-6470,
Volume-4 | Issue-4,
June 2020, pp.1641-
1647, URL:
www.ijtsrd.com/papers/ijtsrd31631.pdf
Copyright © 2020 by author(s) and
International Journal ofTrendinScientific
Research and Development Journal. This
is an Open Access article distributed
under the terms of
the Creative
CommonsAttribution
License (CC BY 4.0)
(http://creativecommons.org/licenses/by
/4.0)
I. INTRODUCTION
Real time object detection is vital for several real-world
vision applications. Recent studies show that certainobjects
like a face can be detected reliably and fairly quickly. The
speed of object detection isn’t fast enough as needed by the
applications. The goal of this paper is to supply a way which
can help the visually impaired person to detect obstacles
through object detection. One essential sense amonghuman
sense is vision and also vision plays an important role
inhuman perception about the encompassing environment.
Hence, thousands of papers are published on these subjects
that propose a selection of computer vision products and
services by developing new electronicaidsfortheblind.This
paper aims to introduce a proposed system that restores a
central function of the visual system which is that the
identification of surrounding objects. Our contribution is to
introduce the thought of a visible substitution system based
on features extractions and matching to recognize/identify
and locate objects in the video in real-time through an
android app.
Computer vision technologies, especially the deep
convolutional neural network, have been rapidly developed
in recent years. It is promising to use state-of-art, computer
vision techniques to assist people with vision loss. The main
aim is to find the possibility of using the sense hearing to
know visual objects. It is not often realized by many people
that we are capable of identifying the spatial location of a
sound source just by hearing it with twoears.Thework aims
to guide blind people through the output of the system by
voice to navigate them through an android app. The
methodology of this work includes Object Extraction,
Feature Extraction, and Object Comparison.
One of the issues faced by visually impaired individuals is to
travel in a corridor. They cannot findeitherthey needtoturn
left or turn right once they reach the end of the corridor and
they get stuck in between the objects if it comes along their
way. When the visually impaired individuals enter the
corridor, they need to find the border of the sidewalk within
the corridor. They use their walking stick to define their
current location. the existing solutions somewhat helped
visually impaired individuals. But all the solutions has its
drawback. And there’s no existingmethodthathelps tosolve
the all basic needs of blind people. All the existing systems
are designed just for a selected purpose.
Although object tracking represents a fundamental problem
in the computer vision community due to the wide range of
related applications (e.g., automatic video surveillance,
wearable assistive devices, robot navigation, structure from
motion, or human-machine interaction), it is still an open
issue of research. Despite significant progress achieved
within the last decade, tremendous challenges still exist in
designing a strong object tracker able to handle important
changes in viewpoint, object motion, light variation, object
occlusion, or background clutter. Through this paper, we
propose a completely unique joint object
detection/recognitionframe-work,calledGODEYE,basedon
computer vision algorithms and offline trained deep
convolutional neural networks. The proposed framework is
IJTSRD31631
International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1642
integrated within a device, dedicated to visually impaired
(VI) users. Here we making an android app for the users.
There are several object detection strategies/algorithms
which will be used here like YOLO[2][21], SSD[11],
RCNN[16], Fast R-CNN[9], Faster R-CNN[10], R-FCN[12].
These algorithms give different rates of performance. So, by
consideringtheimplementationsimplicityandforimproving
performance comparing with other object detection
algorithms SSD gives better performance and simple
implementation.
So, our contribution is an android app for visually impaired
people using the better performing object detection
algorithm with voice assistance and measuring the distance
using an ultrasonic sensor which is helpful for the blinds in
their day-to-day life like staying off from the obstacles and
easily wandering in the new surroundings by the voice
output and the alert signal.
Here we are using the model SSD for object detection. Here
we are employing a trained model SSD MobileNet V1 coco
model, so that we can implement the system godeye easily.
the object detection algorithm used here gives a better
performance than the other existing object detection
algorithm as it follows the bounding box prediction method
for detecting the objects. So we are using a better algorithm
here for detecting the object efficiently and with great
accuracy.
The method that we propose here has the following
specification:
Real time processing & output.
Better efficiency than other existing methods.
Easy to handle.
Alert signal after measuring distance.
The object detection CNN i.e. SSD ensures real-time pro-
cessing and this deep learning model method used here is
much helpful in implementing the system easier than using
other object detection algorithmsandthereforethedetected
object will be given as voice feedback by using TTS[18]. The
distance measurement and alerting also become easy by
using the ultrasonic sensor.
We summarize our contributions as follows:
We propose GODEYE, with an object detection app
based on SSD, it achieves an accurate detection result.
We propose GODEYE, with a distance calculation and
alerting system based on an ultrasonic sensor, it helps
users by alerting from the moving obstacles.
The camera and ultrasonic sensor are attached to the
blind stick. The camera will be connected to the app to
detect the object and provides the voice assistance.
II. PROBLEM DEFENITION
Visually impaired people face many challenges and one
among the main challenges faced by them is self-navigation.
It’s very difficult for them to travel througha corridor.Thisis
one among the main challenge faced by them. They find it
hard to locate the directions i.e. they can’t find whether or
not they need to either turn left or turn right oncetheyreach
the end of the corridor and that they get stuck in between
the objects. While walking through the corridor they should
find the border of the sidewalk. They use a walking stick to
define the situation. Although the present arrangements
helped them yet these arrangements have downsides.
There’s no existing solution that solves all the requirements
for the assistance of a visually impaired person. The existing
solutions designed for a selected purpose. Each solution
solved only a selected problem. Those never solved the
whole problems as an entire.
III. RELATED WORK
NETRA[3], the smart hand glove which acts as an electronic
eye is among the related works. The glove detects the
obstacles within the path ofthevisuallyimpairedperson and
it warns them. It works as an artificial eye for the person.
The glove extracts the text from any image that containstext
and converts that into speech. Another feature of this glove
is that it helps the people to spot the object around them.
The glove sorts out all the issues faced by the people. All the
above tasks are very essential for the safe navigation of
blinds.
NavBelt[4] is another related technology. this is often a
guidance device that uses a mobile robot obstacle avoidance
system. NavBelt is worn by the user like a belt and is
equipped with an array of ultrasonic sensors. This will be
utilized as both primary and secondaryaid.Thisconsistsofa
portable computer, ultrasonic sensors, and stereophonic
imaging technique to process the signal from the ultrasonic
sensors. Acoustic signals are provided via a group of stereo
earphones which guides the user around obstacles or
displays a virtual acoustic panoramic image of the traveler’s
surrounding. Acoustic signals are transmitted in the form of
discrete beeps or continuous sounds. Experimental results
with the NavBelt simulator and a portable prototype shows
that users can travel safely in any unfamiliar and cluttered
environment at speed of up to 0.8m/s.
The walking cane or the guide cane is another related
technology. It detects the obstacles on the way using IR
sensors and produces a buzzing sound as a warning. Using
walking cane the obstacles only up to the knee-level withina
limited range of 2-3 feet can only be detected. The main
drawback is that the buzzer’s sound isn’t audible within the
traffic sounds on the roads[5].
The Voice-enabled smart walking stick is another related
technology. Ultrasonic sensors, Pit sensing unit, Water
sensing unit, GPS module, GSM module Voice synthesizer,
relay, speaker, embeddedsystem,and batteryareusedin the
system. The ultrasonic sensors are used in different sensing
units of the system. If the smart walking stick is madewithat
most accuracy, then it’ll help the blind people to move from
one place to a different without other’s help. This could also
be considered an unrefined way of giving theblinda senseof
vision. The usage of the stick can reduce the visually
impaired from relying on otherfamilymembers,friends,and
guide dogswhilewalkingaround.Theproposedcombination
of various working units makes a real-time system. The
position of the user monitored is the system and also
provides dual feedback. This feedback makes navigation
more safe and secure. Objects or obstaclesaheadoftheusers
are detected using the smart stick and warning voice
messages are sent as feedback instead of vibrations. The
automated room equipment switching incorporated within
the stick are going to be useful while they’re indoor. The
International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1643
advantage of the system lies in the fact that it will prove to
be a low-cost solution to many blind people worldwide[6].
Another related work is a smart navigational shoe. An
electronic kit that is fixed in the shoe can be used by the
blind or visually impaired person. The hardware kitconsists
of four vibrators, one Arduino, Bluetooth connection, and
one battery. The obstacles along the path are sensed by
sensors while moving along the path, if any obstacles
encounter the path this will detected and informs the user.
An android application that can be integratedwith thesmart
navigational shoe is developed here. The starting point and
the destination point are given as input by theuser.Thepath
for navigation is displayed using maps to users. The
application which is integrated with shoes directs the
person. The vibrators will vibrate along the path if an
obstacle is detected. If it’s a left turn that is to be taken
according to the navigation, the vibrator assail the left side
will vibrate. If it’s a right turn that’s to be taken according to
the navigation, the vibrator set on the right side will
vibrate[7].
A smart cap for the blind using the Raspberry Pi system is
another related work that’s easy to understand and
maintain. This technique uses Raspberry pi which is a small
processing device that works as a computer at a
comparatively low cost. Visually impaired people find
difficulties in detecting obstacles while walking in a street.
By making use of Raspberry Pi, the system is intended to
provide artificial vision and object detection, real-time
assistance via GPS. The system consists of Ultrasonic
sensors, GPS module, and therefore the feedback is received
as audio. Voice output works through TTS (text to speech).
The proposed system detects an object around them and
sends feedback in the form of speech that’s warning
messages via earphones. It can also provide navigation to a
selected location through GPS. the general system aims to
provide a low cost, efficient navigation, and obstacle
detection aid for the visually impaired which provides a
sense of artificial vision by providing them with information
about the environmental scenario of the static and dynamic
object around them so that they can walk independently[8].
IV. GODEYE
The system is known as GODEYE. It includes an object
detection app with voiceassistanceandUltrasonic sensor for
distance calculation and alerting. Normally, visually
impaired people use a cane as a guide for them to protect
themselves from obstacles. Most of the surroundingarea are
often covered with the assistance of the cane, especially the
area nearer to their legs like stairs, etc. The system is
designed to provide complete navigation for the user within
the environment around them. It guides the user about
obstacles, also it provides informationabouttheappropriate
obstacle-free path. This technique is an intelligent
monitoring and alert system that involvestheobservation of
real-time activities, with real-time constraints on video
surveillance. The system consists of an android app
connected to the camera attached to the blind’s stick. This
camera detects the obstacles/objects that come along the
path of the visually impaired people or the users and an
ultrasonic sensor attached to the blind stick measures the
distance between the user and therefore the obstacle. The
information is provided through voice outputbytheappand
therefore the sensor gives an alert signal if the obstacle is
extremely near to the user. Fig.1 shows the diagram of the
system.
Fig.1. A block visualization of the system
The main challenge is to get the response in minimum time
and also that the efficiency and accuracy of detecting the
object should be maximum and it should also guide the user
correctly. to overcome these challenges, here we are using
better technologies and algorithms. We are using the SSD
algorithm for better object detection and accuracy and
Ultrasonic sensor for measuring the distance. To alert the
user we are using buzzer which produces a beep sound. On
opening the app the video surveillance begins. The
streaming video is provided as an input to the object
detection algorithm. Usingtheobjectdetectionalgorithm the
obstacles/objects that come along the navigation pathofthe
user are detected. The backend used for object detection is
Tensor Flow. the object detection algorithmgivestheoutput
as a string and therefore the string is given as an input to the
TTS[18]. The TTS provides a voice output to the user. The
architecture diagram of the android app is shown in Fig.2.
Simultaneously the distance calculation between the user
and the obstacle is done using the ultrasonicsensorattached
to the blind stick. For distance measuring the ultrasonic
sensor is attached to Arduino UNO[19], and for the signal
alert, a buzzer is attached to it. The space between user and
object be calculated. A minimum distance is fixed, then the
calculated distance andthe minimumdistanceiscomparedif
the distance between the user and the obstacle is less than
the fixed minimum distance or equal to the fixed minimum
distance the alert system gives a beep sound to the user
which is inbuilt on the stick. The connections of distance
calculation with a buzzer are shown in Fig.3. This how the
GODEYE works.
A. OBJECT DETECTION
The detection of objects within the videos is done using
feature extraction and detection. Feature matching using
invariant features has obtained significant importance with
its application in various recognition problems. These
techniques have helped us to match images regardless of
various geometric and photometric transformations
between images.
The visual substitution system here is based on the
identification of objects around the visually challenged
person using an android App connected to the camera
attached to the blind’s stick. This technique should find the
International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1644
invariant characteristic of objects to viewpoint changes,
provide the recognition, and reduce the complexity of
detection. The method we’ve proposed here is based on the
key-points extraction and matching in the video. A
comparison is formed between the query frame and
database objects to detect the objects in each frame. For
every object detected a string file will be created and stored.
Hence object detected string aregoingtobestoredandgiven
as input to TTS.
The video will be loaded using the real-time camera for
object detection; then feature extraction is done.SSD(Single
Shot Detector) algorithm is employed here for object
detection. When the app is opened the video surveillance
begins. The streaming video is provided as an input to the
object detection algorithm. Using the object detection
algorithm the obstacles/objects that come along the
navigation path of the user are detected. The object
detection algorithm gives the output as a string and
therefore the string will be stored. Then the object detected
string is given TTS for audio/Voice output of the object
detected.
Here for object detection, we are using SSD300. By using
SSD300[20], we only need to takeoneshottodetectmultiple
objects. It achieves a 74.3% mAP at 59 FPS. The network
architecture of SSD300[24][25] is shown in Fig.4. to urge
more accurate detection results, different layers of feature
maps also are browsing the convolution forobjectdetection.
Each convolution will provides a specific number of
bounding boxes for object detection. So, in total SSD300
gives 8732 boxes in total which is more than that of
YOLO[13][14][15]. Due to this much bounding box, object
detection is accurate and efficient. The model usedforobject
detection here is the ssd mobile net v1 coco model. In SSD
the image is divided into grids and each grid cell is
responsible for object detection in that region. By dividing
into grid cells there maybelossoccurswhiledetection.based
on this the loss function of SSD is defined as:
L (x; c; l; g) = 1 (Lconf (x; c) + Lloc(x; l; g))(1)
N
Where L is loss function [20], Lconf is the confidence loss,Lloc
is localization loss, is the weight for the localization loss, l is
predicted box, g is ground truth box,N isthematcheddefault
boxes, c is the multiple class confidences and x is the
indicator for matching defaultboxwiththegroundtruth box.
Based on the above object detection algorithmtechniquethe
object detection is implemented in-app. In this, the input
frame must be 300x300. The SSD300 gives a faster and
accurate results of object detection. The object detection is
the major a part of the GODEYE system for visually impaired
people.
B. TEXT-TO-SPEECH
The Google Text-to-Speech API is employed for getting the
voice output of the detected object. The objects detected by
class predictions in every frame will be a string e.g. “cat”and
it’ll be stored. This string will be given as input to the TTS
converter. The TTS converter will convert the string that got
into the user’s natural language using thegTTspackageused
in the system. The TTS converter will do text analysis to the
string given as input for converting it to the user’s natural
language. After the text will be searched in the speech
database and choose the units compared to that. The speech
generation module that presents within the TTS converter
will make the text/string into speech and given as speech.
This TTS process is completed by a text-to-speech
synthesizer. The general process steps are shown in Fig.5.
The speech made up of the input string will be given as the
output from the android app. this is the main part of the
implemented system. If multiple objects detected from one
frame then based on the accuracy of the object the speech
out will be got on the user’s natural language. The speech
from the TTS[22] is the major output from the app for the
visually impaired person and it’s heard by the user using
earphones or by the phone speaker[18].
C. DISTANCE MEASUREMENT
The ultrasonic sensor is employed here for distance
measurement between the userandtheobject/obstacle[23].
The ultrasonic sensor[19] attached to Arduino UNO are
going to be attached to the blind’s stick. The ultrasonic
sensor one eye produces waves to detect the object, when
the waves hit the object the waves will reflect the opposite
eye of the ultrasonic sensor shown in Fig.6 by that the space
is measured
Fig.2. The system design of Android App.
International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1645
Fig.3. The connections of distance calculation and alert.
between the thing and therefore the user. The Arduino IDE isemployedhereforcodingtheultrasonicsensor.Itusesembedded
C for coding the sensor. within the sensor, we set a threshold distance so that when the obstacle/objectisequal orlessthanthe
threshold distance then an alert signal will be given to the user. The ultrasonic sensor can detect onlythedistancefrom2cmto
400cm of non-contact measurement functionality with a ranging accuracy which will reach up to 3mm. But it’s a sufficient
distance for the user to get alerted. The ultrasonic sensor will update the distance between the user and the obstacle during a
specific interval of time. So that the alert system can alert the user/visually challenged person on reaching the threshold
distance.
The ultrasonic sensor uses sonar waves for echolocation, whichbeabletomeasurethe distances.Theultrasonicsensormodule
has 4 pins- Ground, VCC, Trig, and Echo. The VCC and Ground pins of the ultrasonic sensor need to be connected to the 5 volt
and ground pins on the Arduino Board and the echo and trig pins to any Digital I/O pin on the Arduino Board. The distance
between the sensor and the object is defined as:
distance = speed x time
s = v t (2)
The speed of the sound above water level is 340 m/s or 0.034 cm/µs. From the Echo pin, we’ll get the time duration whichwill
be double because the wave must travel forward and return backward. to get the distance in cm, we need to multiply the
received duration time from the echo pin by 0.034 and divide it by 2. so that thecorrect distancebetweenthesensor andobject
is defined as: here V = 340 m/s or v = 0.034 cm/µs, so,
s = 0:034 t=2 (3)
This is how distance measurement is done by the ultrasonic sensor presentwithinthe system.Basedonthedistancemeasured
the threshold distance will be set and on that basis, the alert signal is given to the user by using buzzer with a beep sound. The
alert will continue if the user is moving closer to the object. the major contribution from the distance measurement system is:
Measure distance between user and objects.
Give alert signal to user based on the distance measured.
V. SPECIFICATIONS
Android Mobile phone, Arduino UNO, Ultrasonic sensor, Buzzer, Earphone, and White cane.
Fig.4. The system design of SSD.[20]
International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1646
Fig.5. The system architecture of Text-To-Speech.[22]
Fig.6. Distance measurement using ultrasonic sensor.
VI. EXPERIMENTAL EVALUATION
By evaluating the experiment here the predictiontimetaken
by the object recognition android app is 70msto500ms.The
distance calculation and cautioning was given elite with a
deferral of 10ms. The threads given in the app makes the
class expectation simpler.
Table 1 shows the comparison of various systems that is
related works and the implemented work. From the
comparison table we can see that the system we proposed
here is effective and performs superior to other/existing
systems. The above comparison table shows that GODEYEis
a very efficient system for blinds.
TABLE I COMPARISON OF VARIOUS SYSTEMS
Systems
Power
Consumption
Range
Response
time
NETRA Low Medium Slow
Nev-Belt Medium Low Slow
Smart Cap Medium Low Fast
Smart Shoe Medium Low Slow
GODEYE Low High Fast
While calculating distance for this experiment the system is
associated with the serial monitor of arduino IDE to check
the outcome exactness and the alarming sign accuracy. At
the point when the distance calculated arrived at the
threshold distance then the alarm sign will begin at that
point onwards with a particular time stretch. It proceeded
until the user or object moves farther. The distance
calculation yield to the serial monitor of arduino IDE is
appeared in Fig.7.
Fig.7. Measured distance from user to object using
ultrasonic sensor
The output of the system proposed hereisanaudiooutputof
the object detected and the alert signal that is beep sound
that occurs if the distance calculated not exactly or
equivalent to the threshold distance. The experimental
evaluation shows that the framework executedisproductive
and superior to the existing system for visually disabled
people.
VII. CONCLUSION AND FUTURE WORK
Here, we implementing a system GODEYE. A system for
visually disabled people. This system is introduced to
overcome the difficulties faced by the visually disabled. Our
system consist of two parts: Object detection with voice
assistance and Distance measurement. SSD300 is used for
objection detection. This strategy is favored due to its
productivity, exactness, and accuracy. This makes theobject
recognition more faster, so the obstacles can be easily
located by the user. Voice feedback is provided after object
detection. This is done by the text-to-speech conversion
using a text-to-speech converter, i.e. text-to-speech API. The
distance between the obstacles and user is calculated using
an ultrasonic sensor joined to the stick. The ultrasonic
sensor is used to calculate the distance. These features are
clubbed in our system. This makes our system efficient and
enables self-navigation for the visually impaired. Wefeature
the presentation of our framework as far as the speed of
recognition with exceptional proficiency in distance
calculation and high accuracy of detection.
In the near future, the framework can be modified with SSD-
MSN technology. Currently, SSD300 is used. With the usage
of SSD- MSN several images frames can be detected.Nowthe
distance calculation is done using the ultrasonic sensor. In
the future, the distance measurement can be advanced by
International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470
@ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1647
measuring the distance using the camera. By includingthese
progressions to our framework the framework transforms
into a high performance and efficient app.
References
[1] Sabin Khader, Meerakrishna M R, Reshma Roy, and
Willson Joseph C, “An Application for Blind using Text
to Speech Conversion,” in HBRP Publication, of Adv.of
Comp. Tech. and its Applications, vol. 3, issue 1, pp 1-6,
2020. Available:
http://doi.org/10.5281/zenodo.3653872
[2] Zuge Chen, Kehe Wu, Minjian Wang and Wei Li, “YOLO
based human action recognition and localization,”
Procedia Comput. Sci., 2018;133:831–8p
[3] N. K. Srivastava and Satyam Singh, “Netra: Smart Hand
Gloves Com-prises Obstacle Detection, Object
Identification OCR Text to Speech Converter for
Blinds,” IEEE Conference Paper, 2018.
[4] S. Shoval, J. Borenstein and Y. Koren, “Auditory
guidance with the Navbelt-a computerized travel aid
for the blind,” IEEE Transactions on Systems, 1998.
[5] A. Sangami, M. Kavithra, K.Rubin and S. Sivaprakasam,
“Obstacle Detection and Location Finding For Blind
People,” IJIRCCE, March 2015.3(2).
[6] Akhila Jose, Greeshma George, MeenuR Nair,Shilpa MJ
and Mr. Munna Basil Mathai,“Voice Enabled Smart
Walking Stick For Visually Impaired,” IJAREEIE, March
2016.5(3).
[7] Saloni Mohanty, Malavika Karunan, Ibtisam Sayyad,
Shlesha Khursade and Prof. B. B. Gite,” Smart Shoe for
Visually Impaired,” IJARCCE,Vol.6,Issue11,November
2017.
[8] Mr. Yash Ghodegaonkar, Miss. Samruddhi Salvi, Mr.
Mahesh Maske and Prof Kalbande M. S.,” Blind
Guidance Wearable using RaspberryPi,”IJIRCCE,Vol.7,
Special Issue 1, June 2019.
[9] R.Girshick, ”Fast R-CNN,” in proc. IEEE Int. Conf.
Comput. Vis., Dec. 2015,pp. 1440-1448.
[10] S. Ren, K. He, R. Girshick, and J. Sun,”Faster R-CNN:
Towards real-time object detection with region
proposal networks,” in Proc. Adv. Neural Inf. Process.
Syst., 2015, pp. 91-99.
[11] W. Liu, D. Angurlov, D. Erhan, C. Szegedy, S. Reed, C.-Y.
Fu and A. C. Berg,” SSD: Single shot multibox detector,”
in Proc. Eur. Conf. Comput. Vis. Cham, Switzerland:
Springer, 2016, pp. 21-37.
[12] J. Dai, Y. Li, K. He and J. Sun,” R-FCN: Object detection
via region-based fullyconvolutional networks,”inProc.
Adv. Neural Inf. Process. Syst., 2016, pp. 379-387.
[13] J. Redmon, S. Divvala, R. Girshick and A. Farhadi,” You
only look once: Unified,” in Proc. IEEE Conf. Comput.
Vis. Pattern Recognit., Jun. 2016, pp. 779-788.
[14] J. Redmon and A. Farhadi,” YOLO9000: Better, faster,
stronger,” in Proc. IEEE Conf. Comput. Vis Pattern
Recognit., Jul. 2017, pp. 7263-7271.
[15] J. Redmon and A. Farhadi,” YOLOv3: An incremental
improvement,” 2018, arXiv: 1804.02767. [Online].
Available: https://arxiv.org/abs/1804.02767
[16] K. He, G. Gkioxari, P. Dollar and R. Girshick,” Mask R-
CNN,” in Proc. IEEE Int. Conf. Comput. Vis., Oct. 2017,
pp. 2961-2969.
[17] Tsung-Yi Lin, Michael Maire, Serge Belongie, James
Hays, Pietro Per-ona, Deva Ramanan, Piotr Dollar, and
C LawrenceZitnick,”Microsoftcoco:Commonobjectsin
context,” InECCV, 2014.
[18] Suhas R. Mache, Manasi R. Baheti and C. Namrata
Mahender,” Re-view on Text-To-Speech Synthesizer,”
International Journal of Advanced Research in
Computer and Communication Engineering, Vol. 4,
2018.
[19] Junez Riyaz,
https://create.arduino.cc/projecthub/Junezriyaz/dista
nce-measurement-using-arduino-ultrasonic-sensor-
409169
[20] Sik-Ho Tsang,
https://towardsdatascience.com/review-ssd-single-
shot-detector-object-detection-851a94607d11
[21] Jason Yip, https://towardsdatascience.com/object-
detection-with-voice-feedback-yolo-v3-gtts-
6ec732dca91
[22] T. Portele and J. Kramer,” Adapting a TTS system to a
reading machine for the blind,” IEEE, Proceeding of
Fourth International Conference on Spoken Language
Processing. ICSLP, Vol 1, 1996.
[23] Darold Wobschall; Ming Zeng and Bhooma
Srinivasaraghavan,” An Ultrasonic/Optical Pulse
Sensor for Precise Distance Measurements,” IEEE,
Sensors for Industry Conference, 2005.
[24] Jonathan Hui, https://medium.com/@jonathan
hui/ssd-object-detection-single-shot-multibox-
detector-for-real-time-processing-9bd8deac0e06
[25] Eddie Forson,
https://towardsdatascience.com/understanding-ssd-
multibox-real-time-object-detection-in-deep-learning-
495ef744fab

More Related Content

What's hot

IRJET- Study on: Advancement of Smartphone Security by using Iris Scan Detection
IRJET- Study on: Advancement of Smartphone Security by using Iris Scan DetectionIRJET- Study on: Advancement of Smartphone Security by using Iris Scan Detection
IRJET- Study on: Advancement of Smartphone Security by using Iris Scan DetectionIRJET Journal
 
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...IRJET Journal
 
Face recognition smart cane using haar-like features and eigenfaces
Face recognition smart cane using haar-like features and eigenfacesFace recognition smart cane using haar-like features and eigenfaces
Face recognition smart cane using haar-like features and eigenfacesTELKOMNIKA JOURNAL
 
Eye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AIEye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AIDr. Amarjeet Singh
 
Ieeepro techno solutions ieee 2014 embedded project talking assistance ab...
Ieeepro techno solutions    ieee  2014 embedded project talking assistance ab...Ieeepro techno solutions    ieee  2014 embedded project talking assistance ab...
Ieeepro techno solutions ieee 2014 embedded project talking assistance ab...srinivasanece7
 
Detection of immovable objects on visually impaired people walking aids
Detection of immovable objects on visually impaired people walking aidsDetection of immovable objects on visually impaired people walking aids
Detection of immovable objects on visually impaired people walking aidsTELKOMNIKA JOURNAL
 
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...EECJOURNAL
 
Iris based Human Identification
Iris based Human IdentificationIris based Human Identification
Iris based Human Identificationdswazalwar
 
Obstacle Detection and Navigation system for Visually Impaired using Smart Shoes
Obstacle Detection and Navigation system for Visually Impaired using Smart ShoesObstacle Detection and Navigation system for Visually Impaired using Smart Shoes
Obstacle Detection and Navigation system for Visually Impaired using Smart ShoesIRJET Journal
 
IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...
IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...
IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...IRJET Journal
 
Hand Gesture for Multiple Applications
Hand Gesture for Multiple ApplicationsHand Gesture for Multiple Applications
Hand Gesture for Multiple ApplicationsIRJET Journal
 
Hand Gesture for Multiple Applications
Hand Gesture for Multiple ApplicationsHand Gesture for Multiple Applications
Hand Gesture for Multiple ApplicationsIRJET Journal
 
IRJET- A Survey on Indoor Navigation for Blind People
IRJET- A Survey on Indoor Navigation for Blind PeopleIRJET- A Survey on Indoor Navigation for Blind People
IRJET- A Survey on Indoor Navigation for Blind PeopleIRJET Journal
 
IRJET- Secure Vault System using Iris Biometrics and PIC Microcontroller
IRJET-  	  Secure Vault System using Iris Biometrics and PIC MicrocontrollerIRJET-  	  Secure Vault System using Iris Biometrics and PIC Microcontroller
IRJET- Secure Vault System using Iris Biometrics and PIC MicrocontrollerIRJET Journal
 
A new perspective of refractive error calculation with mobile application
A new perspective of refractive error calculation with mobile application A new perspective of refractive error calculation with mobile application
A new perspective of refractive error calculation with mobile application IJECEIAES
 
Communication among blind, deaf and dumb People
Communication among blind, deaf and dumb PeopleCommunication among blind, deaf and dumb People
Communication among blind, deaf and dumb PeopleIJAEMSJORNAL
 

What's hot (20)

V4 n2 139
V4 n2 139V4 n2 139
V4 n2 139
 
IRJET- Study on: Advancement of Smartphone Security by using Iris Scan Detection
IRJET- Study on: Advancement of Smartphone Security by using Iris Scan DetectionIRJET- Study on: Advancement of Smartphone Security by using Iris Scan Detection
IRJET- Study on: Advancement of Smartphone Security by using Iris Scan Detection
 
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
IRJET- Review on Raspberry Pi based Assistive Communication System for Blind,...
 
Face recognition smart cane using haar-like features and eigenfaces
Face recognition smart cane using haar-like features and eigenfacesFace recognition smart cane using haar-like features and eigenfaces
Face recognition smart cane using haar-like features and eigenfaces
 
Eye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AIEye(I) Still Know! – An App for the Blind Built using Web and AI
Eye(I) Still Know! – An App for the Blind Built using Web and AI
 
Ieeepro techno solutions ieee 2014 embedded project talking assistance ab...
Ieeepro techno solutions    ieee  2014 embedded project talking assistance ab...Ieeepro techno solutions    ieee  2014 embedded project talking assistance ab...
Ieeepro techno solutions ieee 2014 embedded project talking assistance ab...
 
Detection of immovable objects on visually impaired people walking aids
Detection of immovable objects on visually impaired people walking aidsDetection of immovable objects on visually impaired people walking aids
Detection of immovable objects on visually impaired people walking aids
 
Hands on with Edge AI
Hands on with Edge AIHands on with Edge AI
Hands on with Edge AI
 
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
The use of Eye Gaze Gesture Interaction Artificial Intelligence Techniques fo...
 
Iris based Human Identification
Iris based Human IdentificationIris based Human Identification
Iris based Human Identification
 
Obstacle Detection and Navigation system for Visually Impaired using Smart Shoes
Obstacle Detection and Navigation system for Visually Impaired using Smart ShoesObstacle Detection and Navigation system for Visually Impaired using Smart Shoes
Obstacle Detection and Navigation system for Visually Impaired using Smart Shoes
 
Computer vision ppt
Computer vision pptComputer vision ppt
Computer vision ppt
 
IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...
IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...
IRJET - Using Convolutional Neural Network in Surveillance Videos for Recogni...
 
Hand Gesture for Multiple Applications
Hand Gesture for Multiple ApplicationsHand Gesture for Multiple Applications
Hand Gesture for Multiple Applications
 
Hand Gesture for Multiple Applications
Hand Gesture for Multiple ApplicationsHand Gesture for Multiple Applications
Hand Gesture for Multiple Applications
 
Retina scan
Retina scanRetina scan
Retina scan
 
IRJET- A Survey on Indoor Navigation for Blind People
IRJET- A Survey on Indoor Navigation for Blind PeopleIRJET- A Survey on Indoor Navigation for Blind People
IRJET- A Survey on Indoor Navigation for Blind People
 
IRJET- Secure Vault System using Iris Biometrics and PIC Microcontroller
IRJET-  	  Secure Vault System using Iris Biometrics and PIC MicrocontrollerIRJET-  	  Secure Vault System using Iris Biometrics and PIC Microcontroller
IRJET- Secure Vault System using Iris Biometrics and PIC Microcontroller
 
A new perspective of refractive error calculation with mobile application
A new perspective of refractive error calculation with mobile application A new perspective of refractive error calculation with mobile application
A new perspective of refractive error calculation with mobile application
 
Communication among blind, deaf and dumb People
Communication among blind, deaf and dumb PeopleCommunication among blind, deaf and dumb People
Communication among blind, deaf and dumb People
 

Similar to Godeye An Efficient System for Blinds

SMART BLIND STICK USING VOICE MODULE
SMART BLIND STICK USING VOICE MODULESMART BLIND STICK USING VOICE MODULE
SMART BLIND STICK USING VOICE MODULEIRJET Journal
 
Design and implementation of smart guided glass for visually impaired people
Design and implementation of smart guided glass for visually  impaired peopleDesign and implementation of smart guided glass for visually  impaired people
Design and implementation of smart guided glass for visually impaired peopleIJECEIAES
 
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...IRJET Journal
 
ALTERNATE EYES FOR BLIND advanced wearable for visually impaired people
ALTERNATE EYES FOR BLIND advanced wearable for visually impaired peopleALTERNATE EYES FOR BLIND advanced wearable for visually impaired people
ALTERNATE EYES FOR BLIND advanced wearable for visually impaired peopleIRJET Journal
 
An Assistive System for Visually Impaired People
An Assistive System for Visually Impaired PeopleAn Assistive System for Visually Impaired People
An Assistive System for Visually Impaired PeopleIRJET Journal
 
Sanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance SystemSanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance SystemIRJET Journal
 
Assistance Application for Visually Impaired - VISION
Assistance Application for Visually  Impaired - VISIONAssistance Application for Visually  Impaired - VISION
Assistance Application for Visually Impaired - VISIONIJSRED
 
Real Time Object Detection with Audio Feedback using Yolo v3
Real Time Object Detection with Audio Feedback using Yolo v3Real Time Object Detection with Audio Feedback using Yolo v3
Real Time Object Detection with Audio Feedback using Yolo v3ijtsrd
 
Smart Gloves for Blind
Smart Gloves for BlindSmart Gloves for Blind
Smart Gloves for BlindIRJET Journal
 
IRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
IRJET-Voice Assisted Blind Stick using Ultrasonic SensorIRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
IRJET-Voice Assisted Blind Stick using Ultrasonic SensorIRJET Journal
 
Real Time Translator for Sign Language
Real Time Translator for Sign LanguageReal Time Translator for Sign Language
Real Time Translator for Sign Languageijtsrd
 
A Survey on Smart Devices for Object and Fall Detection
A Survey on Smart Devices for Object and Fall DetectionA Survey on Smart Devices for Object and Fall Detection
A Survey on Smart Devices for Object and Fall DetectionIRJET Journal
 
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfPaper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfSathishRadcliffe
 
A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...
A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...
A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...ijtsrd
 
ARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLE
ARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLEARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLE
ARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLEIRJET Journal
 
Face Mask Detection System Using Artificial Intelligence
Face Mask Detection System Using Artificial IntelligenceFace Mask Detection System Using Artificial Intelligence
Face Mask Detection System Using Artificial IntelligenceIRJET Journal
 
DRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIRED
DRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIREDDRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIRED
DRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIREDIRJET Journal
 
Real-Time Sign Language Detector
Real-Time Sign Language DetectorReal-Time Sign Language Detector
Real-Time Sign Language DetectorIRJET Journal
 
IRJET - Blind Guidance using Smart Cap
IRJET - Blind Guidance using Smart CapIRJET - Blind Guidance using Smart Cap
IRJET - Blind Guidance using Smart CapIRJET Journal
 
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...An assistive model of obstacle detection based on deep learning: YOLOv3 for v...
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...IJECEIAES
 

Similar to Godeye An Efficient System for Blinds (20)

SMART BLIND STICK USING VOICE MODULE
SMART BLIND STICK USING VOICE MODULESMART BLIND STICK USING VOICE MODULE
SMART BLIND STICK USING VOICE MODULE
 
Design and implementation of smart guided glass for visually impaired people
Design and implementation of smart guided glass for visually  impaired peopleDesign and implementation of smart guided glass for visually  impaired people
Design and implementation of smart guided glass for visually impaired people
 
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
INDOOR AND OUTDOOR NAVIGATION ASSISTANCE SYSTEM FOR VISUALLY IMPAIRED PEOPLE ...
 
ALTERNATE EYES FOR BLIND advanced wearable for visually impaired people
ALTERNATE EYES FOR BLIND advanced wearable for visually impaired peopleALTERNATE EYES FOR BLIND advanced wearable for visually impaired people
ALTERNATE EYES FOR BLIND advanced wearable for visually impaired people
 
An Assistive System for Visually Impaired People
An Assistive System for Visually Impaired PeopleAn Assistive System for Visually Impaired People
An Assistive System for Visually Impaired People
 
Sanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance SystemSanjaya: A Blind Assistance System
Sanjaya: A Blind Assistance System
 
Assistance Application for Visually Impaired - VISION
Assistance Application for Visually  Impaired - VISIONAssistance Application for Visually  Impaired - VISION
Assistance Application for Visually Impaired - VISION
 
Real Time Object Detection with Audio Feedback using Yolo v3
Real Time Object Detection with Audio Feedback using Yolo v3Real Time Object Detection with Audio Feedback using Yolo v3
Real Time Object Detection with Audio Feedback using Yolo v3
 
Smart Gloves for Blind
Smart Gloves for BlindSmart Gloves for Blind
Smart Gloves for Blind
 
IRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
IRJET-Voice Assisted Blind Stick using Ultrasonic SensorIRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
IRJET-Voice Assisted Blind Stick using Ultrasonic Sensor
 
Real Time Translator for Sign Language
Real Time Translator for Sign LanguageReal Time Translator for Sign Language
Real Time Translator for Sign Language
 
A Survey on Smart Devices for Object and Fall Detection
A Survey on Smart Devices for Object and Fall DetectionA Survey on Smart Devices for Object and Fall Detection
A Survey on Smart Devices for Object and Fall Detection
 
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdfPaper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
Paper_39-SRAVIP_Smart_Robot_Assistant_for_Visually_Impaired.pdf
 
A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...
A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...
A Deep Neural Framework for Continuous Sign Language Recognition by Iterative...
 
ARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLE
ARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLEARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLE
ARTIFICIAL INTELLIGENCE BASED SMART NAVIGATION SYSTEM FOR BLIND PEOPLE
 
Face Mask Detection System Using Artificial Intelligence
Face Mask Detection System Using Artificial IntelligenceFace Mask Detection System Using Artificial Intelligence
Face Mask Detection System Using Artificial Intelligence
 
DRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIRED
DRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIREDDRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIRED
DRISHTI – A PORTABLE PROTOTYPE FOR VISUALLY IMPAIRED
 
Real-Time Sign Language Detector
Real-Time Sign Language DetectorReal-Time Sign Language Detector
Real-Time Sign Language Detector
 
IRJET - Blind Guidance using Smart Cap
IRJET - Blind Guidance using Smart CapIRJET - Blind Guidance using Smart Cap
IRJET - Blind Guidance using Smart Cap
 
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...An assistive model of obstacle detection based on deep learning: YOLOv3 for v...
An assistive model of obstacle detection based on deep learning: YOLOv3 for v...
 

More from ijtsrd

‘Six Sigma Technique’ A Journey Through its Implementation
‘Six Sigma Technique’ A Journey Through its Implementation‘Six Sigma Technique’ A Journey Through its Implementation
‘Six Sigma Technique’ A Journey Through its Implementationijtsrd
 
Edge Computing in Space Enhancing Data Processing and Communication for Space...
Edge Computing in Space Enhancing Data Processing and Communication for Space...Edge Computing in Space Enhancing Data Processing and Communication for Space...
Edge Computing in Space Enhancing Data Processing and Communication for Space...ijtsrd
 
Dynamics of Communal Politics in 21st Century India Challenges and Prospects
Dynamics of Communal Politics in 21st Century India Challenges and ProspectsDynamics of Communal Politics in 21st Century India Challenges and Prospects
Dynamics of Communal Politics in 21st Century India Challenges and Prospectsijtsrd
 
Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...
Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...
Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...ijtsrd
 
The Impact of Digital Media on the Decentralization of Power and the Erosion ...
The Impact of Digital Media on the Decentralization of Power and the Erosion ...The Impact of Digital Media on the Decentralization of Power and the Erosion ...
The Impact of Digital Media on the Decentralization of Power and the Erosion ...ijtsrd
 
Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...
Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...
Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...ijtsrd
 
Problems and Challenges of Agro Entreprenurship A Study
Problems and Challenges of Agro Entreprenurship A StudyProblems and Challenges of Agro Entreprenurship A Study
Problems and Challenges of Agro Entreprenurship A Studyijtsrd
 
Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...
Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...
Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...ijtsrd
 
The Impact of Educational Background and Professional Training on Human Right...
The Impact of Educational Background and Professional Training on Human Right...The Impact of Educational Background and Professional Training on Human Right...
The Impact of Educational Background and Professional Training on Human Right...ijtsrd
 
A Study on the Effective Teaching Learning Process in English Curriculum at t...
A Study on the Effective Teaching Learning Process in English Curriculum at t...A Study on the Effective Teaching Learning Process in English Curriculum at t...
A Study on the Effective Teaching Learning Process in English Curriculum at t...ijtsrd
 
The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...
The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...
The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...ijtsrd
 
Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...
Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...
Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...ijtsrd
 
Sustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. Sadiku
Sustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. SadikuSustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. Sadiku
Sustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. Sadikuijtsrd
 
Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...
Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...
Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...ijtsrd
 
Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...
Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...
Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...ijtsrd
 
Activating Geospatial Information for Sudans Sustainable Investment Map
Activating Geospatial Information for Sudans Sustainable Investment MapActivating Geospatial Information for Sudans Sustainable Investment Map
Activating Geospatial Information for Sudans Sustainable Investment Mapijtsrd
 
Educational Unity Embracing Diversity for a Stronger Society
Educational Unity Embracing Diversity for a Stronger SocietyEducational Unity Embracing Diversity for a Stronger Society
Educational Unity Embracing Diversity for a Stronger Societyijtsrd
 
Integration of Indian Indigenous Knowledge System in Management Prospects and...
Integration of Indian Indigenous Knowledge System in Management Prospects and...Integration of Indian Indigenous Knowledge System in Management Prospects and...
Integration of Indian Indigenous Knowledge System in Management Prospects and...ijtsrd
 
DeepMask Transforming Face Mask Identification for Better Pandemic Control in...
DeepMask Transforming Face Mask Identification for Better Pandemic Control in...DeepMask Transforming Face Mask Identification for Better Pandemic Control in...
DeepMask Transforming Face Mask Identification for Better Pandemic Control in...ijtsrd
 
Streamlining Data Collection eCRF Design and Machine Learning
Streamlining Data Collection eCRF Design and Machine LearningStreamlining Data Collection eCRF Design and Machine Learning
Streamlining Data Collection eCRF Design and Machine Learningijtsrd
 

More from ijtsrd (20)

‘Six Sigma Technique’ A Journey Through its Implementation
‘Six Sigma Technique’ A Journey Through its Implementation‘Six Sigma Technique’ A Journey Through its Implementation
‘Six Sigma Technique’ A Journey Through its Implementation
 
Edge Computing in Space Enhancing Data Processing and Communication for Space...
Edge Computing in Space Enhancing Data Processing and Communication for Space...Edge Computing in Space Enhancing Data Processing and Communication for Space...
Edge Computing in Space Enhancing Data Processing and Communication for Space...
 
Dynamics of Communal Politics in 21st Century India Challenges and Prospects
Dynamics of Communal Politics in 21st Century India Challenges and ProspectsDynamics of Communal Politics in 21st Century India Challenges and Prospects
Dynamics of Communal Politics in 21st Century India Challenges and Prospects
 
Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...
Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...
Assess Perspective and Knowledge of Healthcare Providers Towards Elehealth in...
 
The Impact of Digital Media on the Decentralization of Power and the Erosion ...
The Impact of Digital Media on the Decentralization of Power and the Erosion ...The Impact of Digital Media on the Decentralization of Power and the Erosion ...
The Impact of Digital Media on the Decentralization of Power and the Erosion ...
 
Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...
Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...
Online Voices, Offline Impact Ambedkars Ideals and Socio Political Inclusion ...
 
Problems and Challenges of Agro Entreprenurship A Study
Problems and Challenges of Agro Entreprenurship A StudyProblems and Challenges of Agro Entreprenurship A Study
Problems and Challenges of Agro Entreprenurship A Study
 
Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...
Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...
Comparative Analysis of Total Corporate Disclosure of Selected IT Companies o...
 
The Impact of Educational Background and Professional Training on Human Right...
The Impact of Educational Background and Professional Training on Human Right...The Impact of Educational Background and Professional Training on Human Right...
The Impact of Educational Background and Professional Training on Human Right...
 
A Study on the Effective Teaching Learning Process in English Curriculum at t...
A Study on the Effective Teaching Learning Process in English Curriculum at t...A Study on the Effective Teaching Learning Process in English Curriculum at t...
A Study on the Effective Teaching Learning Process in English Curriculum at t...
 
The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...
The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...
The Role of Mentoring and Its Influence on the Effectiveness of the Teaching ...
 
Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...
Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...
Design Simulation and Hardware Construction of an Arduino Microcontroller Bas...
 
Sustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. Sadiku
Sustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. SadikuSustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. Sadiku
Sustainable Energy by Paul A. Adekunte | Matthew N. O. Sadiku | Janet O. Sadiku
 
Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...
Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...
Concepts for Sudan Survey Act Implementations Executive Regulations and Stand...
 
Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...
Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...
Towards the Implementation of the Sudan Interpolated Geoid Model Khartoum Sta...
 
Activating Geospatial Information for Sudans Sustainable Investment Map
Activating Geospatial Information for Sudans Sustainable Investment MapActivating Geospatial Information for Sudans Sustainable Investment Map
Activating Geospatial Information for Sudans Sustainable Investment Map
 
Educational Unity Embracing Diversity for a Stronger Society
Educational Unity Embracing Diversity for a Stronger SocietyEducational Unity Embracing Diversity for a Stronger Society
Educational Unity Embracing Diversity for a Stronger Society
 
Integration of Indian Indigenous Knowledge System in Management Prospects and...
Integration of Indian Indigenous Knowledge System in Management Prospects and...Integration of Indian Indigenous Knowledge System in Management Prospects and...
Integration of Indian Indigenous Knowledge System in Management Prospects and...
 
DeepMask Transforming Face Mask Identification for Better Pandemic Control in...
DeepMask Transforming Face Mask Identification for Better Pandemic Control in...DeepMask Transforming Face Mask Identification for Better Pandemic Control in...
DeepMask Transforming Face Mask Identification for Better Pandemic Control in...
 
Streamlining Data Collection eCRF Design and Machine Learning
Streamlining Data Collection eCRF Design and Machine LearningStreamlining Data Collection eCRF Design and Machine Learning
Streamlining Data Collection eCRF Design and Machine Learning
 

Recently uploaded

भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,Virag Sontakke
 
Biting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfBiting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfadityarao40181
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxGaneshChakor2
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptxENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptxAnaBeatriceAblay2
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 
Science lesson Moon for 4th quarter lesson
Science lesson Moon for 4th quarter lessonScience lesson Moon for 4th quarter lesson
Science lesson Moon for 4th quarter lessonJericReyAuditor
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Celine George
 
History Class XII Ch. 3 Kinship, Caste and Class (1).pptx
History Class XII Ch. 3 Kinship, Caste and Class (1).pptxHistory Class XII Ch. 3 Kinship, Caste and Class (1).pptx
History Class XII Ch. 3 Kinship, Caste and Class (1).pptxsocialsciencegdgrohi
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxAvyJaneVismanos
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaVirag Sontakke
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
Hybridoma Technology ( Production , Purification , and Application )
Hybridoma Technology  ( Production , Purification , and Application  ) Hybridoma Technology  ( Production , Purification , and Application  )
Hybridoma Technology ( Production , Purification , and Application ) Sakshi Ghasle
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
Pharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfPharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfMahmoud M. Sallam
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdfssuser54595a
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfSumit Tiwari
 

Recently uploaded (20)

भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
 
Biting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdfBiting mechanism of poisonous snakes.pdf
Biting mechanism of poisonous snakes.pdf
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptx
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptxENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
ENGLISH5 QUARTER4 MODULE1 WEEK1-3 How Visual and Multimedia Elements.pptx
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 
Science lesson Moon for 4th quarter lesson
Science lesson Moon for 4th quarter lessonScience lesson Moon for 4th quarter lesson
Science lesson Moon for 4th quarter lesson
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17
 
History Class XII Ch. 3 Kinship, Caste and Class (1).pptx
History Class XII Ch. 3 Kinship, Caste and Class (1).pptxHistory Class XII Ch. 3 Kinship, Caste and Class (1).pptx
History Class XII Ch. 3 Kinship, Caste and Class (1).pptx
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptx
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of India
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
Hybridoma Technology ( Production , Purification , and Application )
Hybridoma Technology  ( Production , Purification , and Application  ) Hybridoma Technology  ( Production , Purification , and Application  )
Hybridoma Technology ( Production , Purification , and Application )
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
Pharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdfPharmacognosy Flower 3. Compositae 2023.pdf
Pharmacognosy Flower 3. Compositae 2023.pdf
 
9953330565 Low Rate Call Girls In Rohini Delhi NCR
9953330565 Low Rate Call Girls In Rohini  Delhi NCR9953330565 Low Rate Call Girls In Rohini  Delhi NCR
9953330565 Low Rate Call Girls In Rohini Delhi NCR
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
 
Staff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSDStaff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSD
 

Godeye An Efficient System for Blinds

  • 1. International Journal of Trend in Scientific Research and Development (IJTSRD) Volume 4 Issue 4, June 2020 Available Online: www.ijtsrd.com e-ISSN: 2456 – 6470 @ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1641 Godeye: An Efficient System for Blinds Sabin Khader1, Meerakrishna M R1, Reshma Roy1, Willson Joseph C2 1UG Scholar, 2Assistant Professor, 1,2Department of Computer Science, Sahrdaya College of Engineering and Technology, Thrissur, Kerala, India ABSTRACT The technology is growing vastly. Everyone in humanity hassomelimitations. One of those limitations is visual disability. So we are here with a system that helps the visually disabled people. The framework here contains object detection with voice assistance within an app and a hardwarepartattachedto the blinds stick for distance calculation. The app is designed to support the blind person to explore freely anywhere he wants. The working of the framework begins by surveilling the situations around the user and distinguishing them utilizing a camera. The app will then detect the objects present in the input video frame by using the SSD algorithm comparingitwith the trained model. The video captured is partitioned into grids to detect the object/obstacle. In this way, the subtleties of the object detected can be achieved and along with it distance measurement can also be calculatedusing specific algorithms. A Text to Speech (TTS) converter is utilized for changing over the data about the object detected into an audio speech format. The framework/application passes on the scene which the blind peopleisgoingin his/her territorial language with the snap of a catch. The technologiesutilized here makes the framework execution effective. KEYWORDS: visual impairment, android phone, object/obstacle detection, text- to-speech How to cite this paper: Sabin Khader | Meerakrishna M R | Reshma Roy | Willson Joseph C "Godeye: An Efficient System for Blinds" Published in International Journal of Trend in Scientific Research and Development(ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-4, June 2020, pp.1641- 1647, URL: www.ijtsrd.com/papers/ijtsrd31631.pdf Copyright © 2020 by author(s) and International Journal ofTrendinScientific Research and Development Journal. This is an Open Access article distributed under the terms of the Creative CommonsAttribution License (CC BY 4.0) (http://creativecommons.org/licenses/by /4.0) I. INTRODUCTION Real time object detection is vital for several real-world vision applications. Recent studies show that certainobjects like a face can be detected reliably and fairly quickly. The speed of object detection isn’t fast enough as needed by the applications. The goal of this paper is to supply a way which can help the visually impaired person to detect obstacles through object detection. One essential sense amonghuman sense is vision and also vision plays an important role inhuman perception about the encompassing environment. Hence, thousands of papers are published on these subjects that propose a selection of computer vision products and services by developing new electronicaidsfortheblind.This paper aims to introduce a proposed system that restores a central function of the visual system which is that the identification of surrounding objects. Our contribution is to introduce the thought of a visible substitution system based on features extractions and matching to recognize/identify and locate objects in the video in real-time through an android app. Computer vision technologies, especially the deep convolutional neural network, have been rapidly developed in recent years. It is promising to use state-of-art, computer vision techniques to assist people with vision loss. The main aim is to find the possibility of using the sense hearing to know visual objects. It is not often realized by many people that we are capable of identifying the spatial location of a sound source just by hearing it with twoears.Thework aims to guide blind people through the output of the system by voice to navigate them through an android app. The methodology of this work includes Object Extraction, Feature Extraction, and Object Comparison. One of the issues faced by visually impaired individuals is to travel in a corridor. They cannot findeitherthey needtoturn left or turn right once they reach the end of the corridor and they get stuck in between the objects if it comes along their way. When the visually impaired individuals enter the corridor, they need to find the border of the sidewalk within the corridor. They use their walking stick to define their current location. the existing solutions somewhat helped visually impaired individuals. But all the solutions has its drawback. And there’s no existingmethodthathelps tosolve the all basic needs of blind people. All the existing systems are designed just for a selected purpose. Although object tracking represents a fundamental problem in the computer vision community due to the wide range of related applications (e.g., automatic video surveillance, wearable assistive devices, robot navigation, structure from motion, or human-machine interaction), it is still an open issue of research. Despite significant progress achieved within the last decade, tremendous challenges still exist in designing a strong object tracker able to handle important changes in viewpoint, object motion, light variation, object occlusion, or background clutter. Through this paper, we propose a completely unique joint object detection/recognitionframe-work,calledGODEYE,basedon computer vision algorithms and offline trained deep convolutional neural networks. The proposed framework is IJTSRD31631
  • 2. International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470 @ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1642 integrated within a device, dedicated to visually impaired (VI) users. Here we making an android app for the users. There are several object detection strategies/algorithms which will be used here like YOLO[2][21], SSD[11], RCNN[16], Fast R-CNN[9], Faster R-CNN[10], R-FCN[12]. These algorithms give different rates of performance. So, by consideringtheimplementationsimplicityandforimproving performance comparing with other object detection algorithms SSD gives better performance and simple implementation. So, our contribution is an android app for visually impaired people using the better performing object detection algorithm with voice assistance and measuring the distance using an ultrasonic sensor which is helpful for the blinds in their day-to-day life like staying off from the obstacles and easily wandering in the new surroundings by the voice output and the alert signal. Here we are using the model SSD for object detection. Here we are employing a trained model SSD MobileNet V1 coco model, so that we can implement the system godeye easily. the object detection algorithm used here gives a better performance than the other existing object detection algorithm as it follows the bounding box prediction method for detecting the objects. So we are using a better algorithm here for detecting the object efficiently and with great accuracy. The method that we propose here has the following specification: Real time processing & output. Better efficiency than other existing methods. Easy to handle. Alert signal after measuring distance. The object detection CNN i.e. SSD ensures real-time pro- cessing and this deep learning model method used here is much helpful in implementing the system easier than using other object detection algorithmsandthereforethedetected object will be given as voice feedback by using TTS[18]. The distance measurement and alerting also become easy by using the ultrasonic sensor. We summarize our contributions as follows: We propose GODEYE, with an object detection app based on SSD, it achieves an accurate detection result. We propose GODEYE, with a distance calculation and alerting system based on an ultrasonic sensor, it helps users by alerting from the moving obstacles. The camera and ultrasonic sensor are attached to the blind stick. The camera will be connected to the app to detect the object and provides the voice assistance. II. PROBLEM DEFENITION Visually impaired people face many challenges and one among the main challenges faced by them is self-navigation. It’s very difficult for them to travel througha corridor.Thisis one among the main challenge faced by them. They find it hard to locate the directions i.e. they can’t find whether or not they need to either turn left or turn right oncetheyreach the end of the corridor and that they get stuck in between the objects. While walking through the corridor they should find the border of the sidewalk. They use a walking stick to define the situation. Although the present arrangements helped them yet these arrangements have downsides. There’s no existing solution that solves all the requirements for the assistance of a visually impaired person. The existing solutions designed for a selected purpose. Each solution solved only a selected problem. Those never solved the whole problems as an entire. III. RELATED WORK NETRA[3], the smart hand glove which acts as an electronic eye is among the related works. The glove detects the obstacles within the path ofthevisuallyimpairedperson and it warns them. It works as an artificial eye for the person. The glove extracts the text from any image that containstext and converts that into speech. Another feature of this glove is that it helps the people to spot the object around them. The glove sorts out all the issues faced by the people. All the above tasks are very essential for the safe navigation of blinds. NavBelt[4] is another related technology. this is often a guidance device that uses a mobile robot obstacle avoidance system. NavBelt is worn by the user like a belt and is equipped with an array of ultrasonic sensors. This will be utilized as both primary and secondaryaid.Thisconsistsofa portable computer, ultrasonic sensors, and stereophonic imaging technique to process the signal from the ultrasonic sensors. Acoustic signals are provided via a group of stereo earphones which guides the user around obstacles or displays a virtual acoustic panoramic image of the traveler’s surrounding. Acoustic signals are transmitted in the form of discrete beeps or continuous sounds. Experimental results with the NavBelt simulator and a portable prototype shows that users can travel safely in any unfamiliar and cluttered environment at speed of up to 0.8m/s. The walking cane or the guide cane is another related technology. It detects the obstacles on the way using IR sensors and produces a buzzing sound as a warning. Using walking cane the obstacles only up to the knee-level withina limited range of 2-3 feet can only be detected. The main drawback is that the buzzer’s sound isn’t audible within the traffic sounds on the roads[5]. The Voice-enabled smart walking stick is another related technology. Ultrasonic sensors, Pit sensing unit, Water sensing unit, GPS module, GSM module Voice synthesizer, relay, speaker, embeddedsystem,and batteryareusedin the system. The ultrasonic sensors are used in different sensing units of the system. If the smart walking stick is madewithat most accuracy, then it’ll help the blind people to move from one place to a different without other’s help. This could also be considered an unrefined way of giving theblinda senseof vision. The usage of the stick can reduce the visually impaired from relying on otherfamilymembers,friends,and guide dogswhilewalkingaround.Theproposedcombination of various working units makes a real-time system. The position of the user monitored is the system and also provides dual feedback. This feedback makes navigation more safe and secure. Objects or obstaclesaheadoftheusers are detected using the smart stick and warning voice messages are sent as feedback instead of vibrations. The automated room equipment switching incorporated within the stick are going to be useful while they’re indoor. The
  • 3. International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470 @ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1643 advantage of the system lies in the fact that it will prove to be a low-cost solution to many blind people worldwide[6]. Another related work is a smart navigational shoe. An electronic kit that is fixed in the shoe can be used by the blind or visually impaired person. The hardware kitconsists of four vibrators, one Arduino, Bluetooth connection, and one battery. The obstacles along the path are sensed by sensors while moving along the path, if any obstacles encounter the path this will detected and informs the user. An android application that can be integratedwith thesmart navigational shoe is developed here. The starting point and the destination point are given as input by theuser.Thepath for navigation is displayed using maps to users. The application which is integrated with shoes directs the person. The vibrators will vibrate along the path if an obstacle is detected. If it’s a left turn that is to be taken according to the navigation, the vibrator assail the left side will vibrate. If it’s a right turn that’s to be taken according to the navigation, the vibrator set on the right side will vibrate[7]. A smart cap for the blind using the Raspberry Pi system is another related work that’s easy to understand and maintain. This technique uses Raspberry pi which is a small processing device that works as a computer at a comparatively low cost. Visually impaired people find difficulties in detecting obstacles while walking in a street. By making use of Raspberry Pi, the system is intended to provide artificial vision and object detection, real-time assistance via GPS. The system consists of Ultrasonic sensors, GPS module, and therefore the feedback is received as audio. Voice output works through TTS (text to speech). The proposed system detects an object around them and sends feedback in the form of speech that’s warning messages via earphones. It can also provide navigation to a selected location through GPS. the general system aims to provide a low cost, efficient navigation, and obstacle detection aid for the visually impaired which provides a sense of artificial vision by providing them with information about the environmental scenario of the static and dynamic object around them so that they can walk independently[8]. IV. GODEYE The system is known as GODEYE. It includes an object detection app with voiceassistanceandUltrasonic sensor for distance calculation and alerting. Normally, visually impaired people use a cane as a guide for them to protect themselves from obstacles. Most of the surroundingarea are often covered with the assistance of the cane, especially the area nearer to their legs like stairs, etc. The system is designed to provide complete navigation for the user within the environment around them. It guides the user about obstacles, also it provides informationabouttheappropriate obstacle-free path. This technique is an intelligent monitoring and alert system that involvestheobservation of real-time activities, with real-time constraints on video surveillance. The system consists of an android app connected to the camera attached to the blind’s stick. This camera detects the obstacles/objects that come along the path of the visually impaired people or the users and an ultrasonic sensor attached to the blind stick measures the distance between the user and therefore the obstacle. The information is provided through voice outputbytheappand therefore the sensor gives an alert signal if the obstacle is extremely near to the user. Fig.1 shows the diagram of the system. Fig.1. A block visualization of the system The main challenge is to get the response in minimum time and also that the efficiency and accuracy of detecting the object should be maximum and it should also guide the user correctly. to overcome these challenges, here we are using better technologies and algorithms. We are using the SSD algorithm for better object detection and accuracy and Ultrasonic sensor for measuring the distance. To alert the user we are using buzzer which produces a beep sound. On opening the app the video surveillance begins. The streaming video is provided as an input to the object detection algorithm. Usingtheobjectdetectionalgorithm the obstacles/objects that come along the navigation pathofthe user are detected. The backend used for object detection is Tensor Flow. the object detection algorithmgivestheoutput as a string and therefore the string is given as an input to the TTS[18]. The TTS provides a voice output to the user. The architecture diagram of the android app is shown in Fig.2. Simultaneously the distance calculation between the user and the obstacle is done using the ultrasonicsensorattached to the blind stick. For distance measuring the ultrasonic sensor is attached to Arduino UNO[19], and for the signal alert, a buzzer is attached to it. The space between user and object be calculated. A minimum distance is fixed, then the calculated distance andthe minimumdistanceiscomparedif the distance between the user and the obstacle is less than the fixed minimum distance or equal to the fixed minimum distance the alert system gives a beep sound to the user which is inbuilt on the stick. The connections of distance calculation with a buzzer are shown in Fig.3. This how the GODEYE works. A. OBJECT DETECTION The detection of objects within the videos is done using feature extraction and detection. Feature matching using invariant features has obtained significant importance with its application in various recognition problems. These techniques have helped us to match images regardless of various geometric and photometric transformations between images. The visual substitution system here is based on the identification of objects around the visually challenged person using an android App connected to the camera attached to the blind’s stick. This technique should find the
  • 4. International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470 @ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1644 invariant characteristic of objects to viewpoint changes, provide the recognition, and reduce the complexity of detection. The method we’ve proposed here is based on the key-points extraction and matching in the video. A comparison is formed between the query frame and database objects to detect the objects in each frame. For every object detected a string file will be created and stored. Hence object detected string aregoingtobestoredandgiven as input to TTS. The video will be loaded using the real-time camera for object detection; then feature extraction is done.SSD(Single Shot Detector) algorithm is employed here for object detection. When the app is opened the video surveillance begins. The streaming video is provided as an input to the object detection algorithm. Using the object detection algorithm the obstacles/objects that come along the navigation path of the user are detected. The object detection algorithm gives the output as a string and therefore the string will be stored. Then the object detected string is given TTS for audio/Voice output of the object detected. Here for object detection, we are using SSD300. By using SSD300[20], we only need to takeoneshottodetectmultiple objects. It achieves a 74.3% mAP at 59 FPS. The network architecture of SSD300[24][25] is shown in Fig.4. to urge more accurate detection results, different layers of feature maps also are browsing the convolution forobjectdetection. Each convolution will provides a specific number of bounding boxes for object detection. So, in total SSD300 gives 8732 boxes in total which is more than that of YOLO[13][14][15]. Due to this much bounding box, object detection is accurate and efficient. The model usedforobject detection here is the ssd mobile net v1 coco model. In SSD the image is divided into grids and each grid cell is responsible for object detection in that region. By dividing into grid cells there maybelossoccurswhiledetection.based on this the loss function of SSD is defined as: L (x; c; l; g) = 1 (Lconf (x; c) + Lloc(x; l; g))(1) N Where L is loss function [20], Lconf is the confidence loss,Lloc is localization loss, is the weight for the localization loss, l is predicted box, g is ground truth box,N isthematcheddefault boxes, c is the multiple class confidences and x is the indicator for matching defaultboxwiththegroundtruth box. Based on the above object detection algorithmtechniquethe object detection is implemented in-app. In this, the input frame must be 300x300. The SSD300 gives a faster and accurate results of object detection. The object detection is the major a part of the GODEYE system for visually impaired people. B. TEXT-TO-SPEECH The Google Text-to-Speech API is employed for getting the voice output of the detected object. The objects detected by class predictions in every frame will be a string e.g. “cat”and it’ll be stored. This string will be given as input to the TTS converter. The TTS converter will convert the string that got into the user’s natural language using thegTTspackageused in the system. The TTS converter will do text analysis to the string given as input for converting it to the user’s natural language. After the text will be searched in the speech database and choose the units compared to that. The speech generation module that presents within the TTS converter will make the text/string into speech and given as speech. This TTS process is completed by a text-to-speech synthesizer. The general process steps are shown in Fig.5. The speech made up of the input string will be given as the output from the android app. this is the main part of the implemented system. If multiple objects detected from one frame then based on the accuracy of the object the speech out will be got on the user’s natural language. The speech from the TTS[22] is the major output from the app for the visually impaired person and it’s heard by the user using earphones or by the phone speaker[18]. C. DISTANCE MEASUREMENT The ultrasonic sensor is employed here for distance measurement between the userandtheobject/obstacle[23]. The ultrasonic sensor[19] attached to Arduino UNO are going to be attached to the blind’s stick. The ultrasonic sensor one eye produces waves to detect the object, when the waves hit the object the waves will reflect the opposite eye of the ultrasonic sensor shown in Fig.6 by that the space is measured Fig.2. The system design of Android App.
  • 5. International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470 @ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1645 Fig.3. The connections of distance calculation and alert. between the thing and therefore the user. The Arduino IDE isemployedhereforcodingtheultrasonicsensor.Itusesembedded C for coding the sensor. within the sensor, we set a threshold distance so that when the obstacle/objectisequal orlessthanthe threshold distance then an alert signal will be given to the user. The ultrasonic sensor can detect onlythedistancefrom2cmto 400cm of non-contact measurement functionality with a ranging accuracy which will reach up to 3mm. But it’s a sufficient distance for the user to get alerted. The ultrasonic sensor will update the distance between the user and the obstacle during a specific interval of time. So that the alert system can alert the user/visually challenged person on reaching the threshold distance. The ultrasonic sensor uses sonar waves for echolocation, whichbeabletomeasurethe distances.Theultrasonicsensormodule has 4 pins- Ground, VCC, Trig, and Echo. The VCC and Ground pins of the ultrasonic sensor need to be connected to the 5 volt and ground pins on the Arduino Board and the echo and trig pins to any Digital I/O pin on the Arduino Board. The distance between the sensor and the object is defined as: distance = speed x time s = v t (2) The speed of the sound above water level is 340 m/s or 0.034 cm/µs. From the Echo pin, we’ll get the time duration whichwill be double because the wave must travel forward and return backward. to get the distance in cm, we need to multiply the received duration time from the echo pin by 0.034 and divide it by 2. so that thecorrect distancebetweenthesensor andobject is defined as: here V = 340 m/s or v = 0.034 cm/µs, so, s = 0:034 t=2 (3) This is how distance measurement is done by the ultrasonic sensor presentwithinthe system.Basedonthedistancemeasured the threshold distance will be set and on that basis, the alert signal is given to the user by using buzzer with a beep sound. The alert will continue if the user is moving closer to the object. the major contribution from the distance measurement system is: Measure distance between user and objects. Give alert signal to user based on the distance measured. V. SPECIFICATIONS Android Mobile phone, Arduino UNO, Ultrasonic sensor, Buzzer, Earphone, and White cane. Fig.4. The system design of SSD.[20]
  • 6. International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470 @ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1646 Fig.5. The system architecture of Text-To-Speech.[22] Fig.6. Distance measurement using ultrasonic sensor. VI. EXPERIMENTAL EVALUATION By evaluating the experiment here the predictiontimetaken by the object recognition android app is 70msto500ms.The distance calculation and cautioning was given elite with a deferral of 10ms. The threads given in the app makes the class expectation simpler. Table 1 shows the comparison of various systems that is related works and the implemented work. From the comparison table we can see that the system we proposed here is effective and performs superior to other/existing systems. The above comparison table shows that GODEYEis a very efficient system for blinds. TABLE I COMPARISON OF VARIOUS SYSTEMS Systems Power Consumption Range Response time NETRA Low Medium Slow Nev-Belt Medium Low Slow Smart Cap Medium Low Fast Smart Shoe Medium Low Slow GODEYE Low High Fast While calculating distance for this experiment the system is associated with the serial monitor of arduino IDE to check the outcome exactness and the alarming sign accuracy. At the point when the distance calculated arrived at the threshold distance then the alarm sign will begin at that point onwards with a particular time stretch. It proceeded until the user or object moves farther. The distance calculation yield to the serial monitor of arduino IDE is appeared in Fig.7. Fig.7. Measured distance from user to object using ultrasonic sensor The output of the system proposed hereisanaudiooutputof the object detected and the alert signal that is beep sound that occurs if the distance calculated not exactly or equivalent to the threshold distance. The experimental evaluation shows that the framework executedisproductive and superior to the existing system for visually disabled people. VII. CONCLUSION AND FUTURE WORK Here, we implementing a system GODEYE. A system for visually disabled people. This system is introduced to overcome the difficulties faced by the visually disabled. Our system consist of two parts: Object detection with voice assistance and Distance measurement. SSD300 is used for objection detection. This strategy is favored due to its productivity, exactness, and accuracy. This makes theobject recognition more faster, so the obstacles can be easily located by the user. Voice feedback is provided after object detection. This is done by the text-to-speech conversion using a text-to-speech converter, i.e. text-to-speech API. The distance between the obstacles and user is calculated using an ultrasonic sensor joined to the stick. The ultrasonic sensor is used to calculate the distance. These features are clubbed in our system. This makes our system efficient and enables self-navigation for the visually impaired. Wefeature the presentation of our framework as far as the speed of recognition with exceptional proficiency in distance calculation and high accuracy of detection. In the near future, the framework can be modified with SSD- MSN technology. Currently, SSD300 is used. With the usage of SSD- MSN several images frames can be detected.Nowthe distance calculation is done using the ultrasonic sensor. In the future, the distance measurement can be advanced by
  • 7. International Journal of Trend in Scientific Research and Development (IJTSRD) @ www.ijtsrd.com eISSN: 2456-6470 @ IJTSRD | Unique Paper ID – IJTSRD31631 | Volume – 4 | Issue – 4 | May-June 2020 Page 1647 measuring the distance using the camera. By includingthese progressions to our framework the framework transforms into a high performance and efficient app. References [1] Sabin Khader, Meerakrishna M R, Reshma Roy, and Willson Joseph C, “An Application for Blind using Text to Speech Conversion,” in HBRP Publication, of Adv.of Comp. Tech. and its Applications, vol. 3, issue 1, pp 1-6, 2020. Available: http://doi.org/10.5281/zenodo.3653872 [2] Zuge Chen, Kehe Wu, Minjian Wang and Wei Li, “YOLO based human action recognition and localization,” Procedia Comput. Sci., 2018;133:831–8p [3] N. K. Srivastava and Satyam Singh, “Netra: Smart Hand Gloves Com-prises Obstacle Detection, Object Identification OCR Text to Speech Converter for Blinds,” IEEE Conference Paper, 2018. [4] S. Shoval, J. Borenstein and Y. Koren, “Auditory guidance with the Navbelt-a computerized travel aid for the blind,” IEEE Transactions on Systems, 1998. [5] A. Sangami, M. Kavithra, K.Rubin and S. Sivaprakasam, “Obstacle Detection and Location Finding For Blind People,” IJIRCCE, March 2015.3(2). [6] Akhila Jose, Greeshma George, MeenuR Nair,Shilpa MJ and Mr. Munna Basil Mathai,“Voice Enabled Smart Walking Stick For Visually Impaired,” IJAREEIE, March 2016.5(3). [7] Saloni Mohanty, Malavika Karunan, Ibtisam Sayyad, Shlesha Khursade and Prof. B. B. Gite,” Smart Shoe for Visually Impaired,” IJARCCE,Vol.6,Issue11,November 2017. [8] Mr. Yash Ghodegaonkar, Miss. Samruddhi Salvi, Mr. Mahesh Maske and Prof Kalbande M. S.,” Blind Guidance Wearable using RaspberryPi,”IJIRCCE,Vol.7, Special Issue 1, June 2019. [9] R.Girshick, ”Fast R-CNN,” in proc. IEEE Int. Conf. Comput. Vis., Dec. 2015,pp. 1440-1448. [10] S. Ren, K. He, R. Girshick, and J. Sun,”Faster R-CNN: Towards real-time object detection with region proposal networks,” in Proc. Adv. Neural Inf. Process. Syst., 2015, pp. 91-99. [11] W. Liu, D. Angurlov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu and A. C. Berg,” SSD: Single shot multibox detector,” in Proc. Eur. Conf. Comput. Vis. Cham, Switzerland: Springer, 2016, pp. 21-37. [12] J. Dai, Y. Li, K. He and J. Sun,” R-FCN: Object detection via region-based fullyconvolutional networks,”inProc. Adv. Neural Inf. Process. Syst., 2016, pp. 379-387. [13] J. Redmon, S. Divvala, R. Girshick and A. Farhadi,” You only look once: Unified,” in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2016, pp. 779-788. [14] J. Redmon and A. Farhadi,” YOLO9000: Better, faster, stronger,” in Proc. IEEE Conf. Comput. Vis Pattern Recognit., Jul. 2017, pp. 7263-7271. [15] J. Redmon and A. Farhadi,” YOLOv3: An incremental improvement,” 2018, arXiv: 1804.02767. [Online]. Available: https://arxiv.org/abs/1804.02767 [16] K. He, G. Gkioxari, P. Dollar and R. Girshick,” Mask R- CNN,” in Proc. IEEE Int. Conf. Comput. Vis., Oct. 2017, pp. 2961-2969. [17] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Per-ona, Deva Ramanan, Piotr Dollar, and C LawrenceZitnick,”Microsoftcoco:Commonobjectsin context,” InECCV, 2014. [18] Suhas R. Mache, Manasi R. Baheti and C. Namrata Mahender,” Re-view on Text-To-Speech Synthesizer,” International Journal of Advanced Research in Computer and Communication Engineering, Vol. 4, 2018. [19] Junez Riyaz, https://create.arduino.cc/projecthub/Junezriyaz/dista nce-measurement-using-arduino-ultrasonic-sensor- 409169 [20] Sik-Ho Tsang, https://towardsdatascience.com/review-ssd-single- shot-detector-object-detection-851a94607d11 [21] Jason Yip, https://towardsdatascience.com/object- detection-with-voice-feedback-yolo-v3-gtts- 6ec732dca91 [22] T. Portele and J. Kramer,” Adapting a TTS system to a reading machine for the blind,” IEEE, Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP, Vol 1, 1996. [23] Darold Wobschall; Ming Zeng and Bhooma Srinivasaraghavan,” An Ultrasonic/Optical Pulse Sensor for Precise Distance Measurements,” IEEE, Sensors for Industry Conference, 2005. [24] Jonathan Hui, https://medium.com/@jonathan hui/ssd-object-detection-single-shot-multibox- detector-for-real-time-processing-9bd8deac0e06 [25] Eddie Forson, https://towardsdatascience.com/understanding-ssd- multibox-real-time-object-detection-in-deep-learning- 495ef744fab