SlideShare a Scribd company logo
1 of 5
Download to read offline
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2633
Survey on Simulation of Self-Driving Cars using Supervised and
Reinforcement Learning Models
Athul Dev1, Roshni K S2
1 Student, Department of Computer Science & Engineering, Nitte Meenakshi Institute of Technology, India
2 Student, Dept. of Electrical & Electronics Engineering, Sir. M Vishveshvaraya Institute of Technology, India
-----------------------------------------------------------------------------***---------------------------------------------------------------------------
Abstract - A self-driving system is the heart, soul and futureof
the automobile industry and there is a lot of techniques and
technology involved in the accomplishment of a self-driving
car. Since the system has to be deployed in the real world and
which is supposed to take multiple actions and sudden
reactions in real time, the system should be tuned to a very
high degree of precision. In this paper we share our
observations with respect to tuning and tweaking most of the
hyper parameters and the architecture in a generated
environment for both supervised and reinforcement learning
techniques. And how different a supervised learning model is
to a reinforcement learning model with respect to the self-
driving car. We use the concepts of Deep Learning, Deep
Neural Networks, Convolutional Networks and Deep Q
learning in order to configure the brain of our car to perform
certain actions.
Key Words: Machine Learning, Supervised Learning,
Reinforcement Learning, Q Learning, Convolutional
Neural Networks, Deep Learning, Markov Decision
Process.
1.INTRODUCTION
From the perspective of a self-driving car, we have mainly
considered the motion, the movement and the overall
physics of the car. In the supervised model wehavemadethe
prediction of the car’s position, its steering angle and the
speed of the car in the environment based on how we drove
the car in that particular environment, so on the basis of
certain frames of images encountered, the required steering
angle and speed in fed back to the car as it’s input. That is
image is the input and the steering angle with the speed,
throttle and brake is the output. So in this case we are not
considering any sorts of external obstacles like another
vehicle, a barricade and so on. Whereasin the reinforcement
model we have considered the obstacles by giving it a
negative reward when it encounters and obstacles or
barriers, due to which it tends to avoid these obstacles. We
have tuned various hyper parameters like the learning rate,
living penalty, rewards, the activation functions and the
network architecture. And the testing for both is done for
various time intervals with its related accuracy value.
The combination of the environments or the learning
techniquescan be made possible which is the futurescopeof
this paper.
2. SUPERVISED LEARNING MODEL
Supervised Learning is the machine learning taskoflearning
a function that mapsan input to an output based on example
input-output pairs. It infers a function from labeled training
data consisting of a set of training examples.
In this case the imagesrecorded from the car when riddenis
the input data and the steering angle is the target label. And
this supervised model in general is a technique that was
implemented by the [1]NVIDIA’sEnd-to-End Deep Learning
for Self-Driving Cars. We have used the concept and
modelled a system with different activation functions and
different convolutional layers.
2.1 Implementation
The implementation is done in three main stages/phases:
a. Data Generation Phase
b. Training Phase
c. Testing Phase
2.1.1 Data Generation Phase
For the data generation part, where the data collectedhereis
the imagesfrom the camera(s), which is nothingbuttheroad
that the car sees to its front and various physics variables
likeitsspeed, throttle, brake and steering. In order to getthe
data and for the environment we have used Udacity’s open
source version of the car simulator which comes with two
pre-loaded environments and an efficient way to train and
test the model.
In order to generate the data we select the training mode of
the open source simulator which brings up a car with three
camera’s mounted on it, towards the left, right and center of
the car. So we start moving the car with the keyboard inputs
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2634
or with a joystick for smother transitions,wecannowrecord
this movement of the car which is controlled by us(ahuman
driver) and this action will be mimicked by ourmodel in the
form of mapping images to actions.
This data can then be saved into a comma separated value
file, which has the images from the camera’s pointing to left,
right and center of the car with the image’s absolute path,
the speed, throttle, brake and the steering values
2.1.2 Training Phase
The training process is inspired by the [1] NVIDIA’s End-to-
End Deep Learning for Self-Driving Cars, they have made use
of an actual car fitted with three camerason itsleft, right and
center and had a human driver drive the car for 72hrs,
recording the images from the cameras along with the car’s
speed, throttle, brake and the steering angle values.
We have mimicked the above process using the Udacity’s
open source version of the car simulator, by making a user
(human driver) drive the car in the simulator using a
keyboardfor various lapsand different intervalsoftime.And
have recorded it with respect to the number of laps and
various time intervals which is shown in the Table 1.
Table -1: Accuracy with respect to the no. of laps.
Number of Laps Time taken to
train the model
Accuracy of the
model
1 15mins 20%
3 50mins 45%
5 3hrs 70%
10 8hrs 85%
In the model, we have used Keras for our deep learning
framework with a sequential model. We start with doing
image normalization and adding a 10 layer convolutional
network layer with 5 fully connected layers to flatten the
data. We have tried the model using various activation
functions like the sigmoid activation function, rectifiedlinear
activation function and the exponential linearunit activation
function, and found exponential linearunitthebestactivation
function for this scenario as it takes care of the vanishing
gradient problem.
Table -2: Accuracy with respect to the activation functions
Activation Function Accuracy of the model
Sigmoid 70-85%
Rectified Linear Unit 80-90%
Exponential Linear Unit 85-95%
Now since the setup is ready, we try to find the data loss or
the error that is by finding the difference between the
predicted steering command and the actual steering
command from the training data. Once we determine the
mean of sumof squared error,that isthenbackpropagatedto
the convolutional neural network which in turn produces a
network computed steering value. And for optimization we
have used Adam optimizer that does our Gradient decent
function.
2.1.3 Testing Phase
A server-client structure is used for the testing, where the
server is the open source simulator and the client, our
python program.
Variousdependencies and librarieslikesocketiowhichhelps
us work in terms of piping commands in real time using
event handlers, Pillow for image manipulation and flask
which is a web framework.
For testing, we start our simulator and then pass the model
that was trained in the training phase to our testing script
and simultaneously activate the autonomous mode in the
simulator to which the program connects and outputs the
required steering angle, speed, throttle and brake values to
the autonomous car.
Here the prediction of the steering angle with respect to the
image from the camerasis sent to the server and Thusgiving
us a self-driving car in a simulated environment.
Fig -1: Testing the model
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2635
3. REINFORCEMENT LEARNING MODEL
Reinforcement learning is a type of Machine Learning
algorithms which allows software agents and machines to
automatically determine the ideal behavior within a specific
context, to maximize its performance.
Reinforcement algorithms are not given explicit goals;
instead, they are forced to learn these optimal goals by trial
and error.
Here we are makinguse ofthe sole conceptsofreinforcement
learning like theBellman Equation,MarkovDecisionProcess,
and Q Learning which is the base for Deep Q Learning
Networks.
3.1 Concepts
3.1.1 Bellman Equation
Say we have a state s, an action taken by the agent as a, the
reward R which the agent gets once it gets into a particular
state and Ὑ (gamma) the discount.
The bellman equation assigns a value to a statebygettingthe
maximum of all the actions possible with respect to the sum
of the rewards that the agent gets in the current state for a
certain action and the value of the following stateorthestate
that the agent would end up in from the current state.
3.1.2 Markov Decision Process
The Markov Decision Process takes care of the non-
deterministic search problem. That is the agent is bound the
take any action in order to find itsoptimal policy or pathand
the actions it might take may not be in a deterministic way,
thus we have a probability term that is multiplied to the
value of the following state.
3.1.3 Living Penalty
Generally an agent gets a reward say in our case a positive
reward when it reaches its destinations and a negative
reward when it goes off road, but a living penalty is a reward
that is given to an agent throughout in its exploration phase
so as to generally find an optimal path. The policy or the
mapping changes with respect to the living penalty value
that is assigned to the agent in its environment.
3.1.4 Q Learning
Q-learning is a simple way for agents to learn how to act
optimally in controlled Markovian domains.Itamountstoan
incremental method for dynamic programming which
imposes limited computational demands. It works by
successively improving its evaluations of the quality of
particular actions at particular states.
The value of the states are replaced by the Q functions or
rather the Quality factor of the action in the respective
states.
3.2 Design and Environment
We have made use of Kivy which is an open source Python
library for development of our self-driving car interface or
the environment and for the mapping purposes.
In the environment we have a car that goes from one
destination to another, where in this casethedestinationsare
predefined. Asit movesitis encounterstomanybarriersand
the car overcomes or surpasses these barriers and continue
its route using the idea of Deep Q learning or in general
Reinforcement learning technique(s).
The user is allowed to place or insert the barriers in the
course of the car and the car surpasses these barriers in real
time with minimal fault and training. The learning, with
respect to this scenario is very fast as compared to the
previously mentioned supervised learning model.
The possible actions that the car can take here is move left,
right or straight. And the angle of the rotation to its left or
right is specified as 10. The car is equipped with three
sensors which are to its left, right and center and these
sensors detect the barriers and avoids them by taking the
opposite direction of the sensor that detects the barriers.
We also configure the rewards that the car gets when it hits
the barrier as-1whichis theworstreward, the livingpenalty
reward as -0.4, if it reaches its optimal path that is a path
better than the previousone in termsof the distance taken it
gets a reward of 0.2, a reward of -1 if it hits the environment
window or the limit and a reward of 1 if it reaches its
destination.
3.3 Implementation
In the reinforcement learning there is no need of explicit
data generation because the data is obtained as the agent is
exploring, and the training and testing happens
simultaneously.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2636
We are extensively using PyTorch deep learning framework
for creating the neural network architecture. The neural
network architecture is comprising of 2 fully connected
layers in a linear manner with 30 nodes in the hidden layer,
the first fully connected layer is with input and the hidden
layer and the next is between the hidden layer and the
number of actions the car can take.
Next we start building the feed forward network, for this we
apply an activation function toour first fully connectedlayer
and generate the Q values by passing the output of the
activation function to the second fully connected layer. And
return the Q values. This architecture can be improved by
tweaking the number of nodesin the hiddenlayersandusing
different activation functions.
Then we continue to implement an experience reply for the
agent as the Markov Decision process or the Q learning
algorithm works with respect to analyzing a seriesof events
and the consecutive states, that is the current state and the
following state. The correlation among these statesaremore
and the network will not be learning very well. So inorderto
avoid this we use experience reply which considers say
about 100-1000 statesand keepsthem in the memoryasthe
series of events so that an efficient decision can be taken by
the agent and these bulk states are treated in various
batches.
Finally we build our Deep Q learning model by passing
various hyper parameterslikethe discount factor (Gamma),
the model itself from the neural network architecture, the
reply memory and the optimizer to do stochastic gradient
decent where in this case an Adam Optimizer is used with
the learning rate of 0.005 which gave the best result.Wealso
determine the right action for the car to take that is whether
to go left, right or straight using a softmax layer. Softmax
gives a large probability to highest Q value. Softmax also
takes care of the exploration of the environmentbytheagent
really well. The learning is done in batches with respect to
the reply memory technique. While learning weconsiderthe
maximum of the Q values of the next state as well as the
reward obtained in the current state. We then calculate the
loss using the smooth l1 loss (Huber loss - since it improves
the Q learning) and pass the loss error to the optimizer todo
the stochastic gradient decent and update the weights using
the back propagation. Eventually an update is done with
respect to the action and the state when the agent gets intoa
new state and a connection is made between the brain (the
DQN logic) and the map function which is defined earlier in
the section 3.2 Design and Environment.
3.4 Training and Testing
The training and testing can be visualized once we run the
environment and the car, since the Deep Q Network (DNQ)
logic is linked to the mapping part we need not explicitly
pass any arguments to link them at the time of execution.
The car starts exploring the environment for some time and
once it reachesitsdestination it gets a positiverewardandit
looksfor itsnext destination checkpoint andcontinuestoget
positive reward from reaching checkpoints and finds an
optimal path to do the same. But when we start placing
barriers in the environment and when the car encounters
these barriers it receives a negative reward, so in its next
encounter it evades the barrier and becomes much efficient
to its previous transition.
The best part about this system is that the barriers or the
obstacles can be randomly anytime and anywhere on the
environment and the car or the agent avoidsor evadesthese
barriers in real time.
Fig -2: Bird’s view of the environment
Thus, the agent updates the rewards and the scores
continuously andtakesthe next actions with respect tothese
updated rewards, scores and the Q values.
This gives rise to a self-driving car using the reinforcement
learning technique.
Chart -1: Behavior with respect to the time and the reward
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2637
4. SUPERVISED V/S REINFORCEMENT MODEL
In the supervised learning model the mapping between the
images to the car’s attributes like speed, steering angle,
throttle and brake is done, so whenever the car/ model gets
a new image it will act accordingly whereasinreinforcement
learning model the car explores the environment at first for
some time as a part of training and with respect to the
rewards it gets, it moves in the environment taking various
decisions like going left, right or straight.
Supervised learning model needs a lot of data and
predefined training for a long time only then it can be
implement accurately whereas reinforcement learning
adapts quickly with less training and data.
Reinforcement learning model in this scenario can easily
evade or surpass any obstacle thrown to it by the user
anytime, anywhere whereas the supervised learning model
fails to do so.
Supervised learning model can easily facilitate a large
number of features and a high accuracy can be obtained but
we will have to use a powerful enough computerthat run on
multiple graphics processing unit(s).
5. CONCLUSIONS
In this paper we have given an overview of the two different
models that is the supervised learning model and the
reinforcement learning model with respect to a self-driving
car. And how each model behaves when the hyper
parameters are tweaked, variations in the activation
functions is made and how long the model is trained.
The model can be furthered improvised by changing the
environments and the architecture, and the implementation
in general can be improved by combining the learning
models and further tweaking the hyper parameters or
adding more layers to the architecture.
ACKNOWLEDGEMENT
We thank our family and friends for their priceless support
and encouragement.
REFERENCES
[1] NVIDIA’s End-to-End Deep Learning for Self-Driving
Cars.
[2] The Theory of Dynamic Programming By Richard
Bellman (1954)
[3] Markov Decision Processes:ConceptsandAlgorithmsBy
Martijn van Otterlo (2009)
[4] Learning to Predict by the Methods of Temporal
Differences BY Richard Sutton (1988)
[5] Reinforcement Learning I: Introduction By Richard
Sutton.
[6] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. R.
Howard, W. Hubbard and L. D. Jackel. Backprop-agation
applied to handwritten zip code recognition.
BIOGRAPHIES
Athul Dev
Machine Learning and Data Science
enthusiast.
Experiences: Software Developer
Intern at Power Grid Corporation of
India, Machine Learning Intern at
Latent Talent Technologies and Curl
Analytics, Data Analyst Intern at CHR
Solutions. Interested in Big Data
Analytics, AI, ML and web
development.
Roshni K S
IoT, embedded systems,
microcontrollers,matlabandcomputer
programming enthusiast.
Experiences: Software Developer
Intern at Power Grid Corporation of
India, Power System Analyst Intern at
KPTCL.

More Related Content

Similar to IRJET-Survey on Simulation of Self-Driving Cars using Supervised and Reinforcement Learning Models

IRJET- Self-Driving Cars: Automation Testing using Udacity Simulator
IRJET- Self-Driving Cars: Automation Testing using Udacity SimulatorIRJET- Self-Driving Cars: Automation Testing using Udacity Simulator
IRJET- Self-Driving Cars: Automation Testing using Udacity SimulatorIRJET Journal
 
IRJET - Steering Wheel Angle Prediction for Self-Driving Cars
IRJET - Steering Wheel Angle Prediction for Self-Driving CarsIRJET - Steering Wheel Angle Prediction for Self-Driving Cars
IRJET - Steering Wheel Angle Prediction for Self-Driving CarsIRJET Journal
 
Self-Driving Car to Drive Autonomously using Image Processing and Deep Learning
Self-Driving Car to Drive Autonomously using Image Processing and Deep LearningSelf-Driving Car to Drive Autonomously using Image Processing and Deep Learning
Self-Driving Car to Drive Autonomously using Image Processing and Deep LearningIRJET Journal
 
Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...
Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...
Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...IRJET Journal
 
AIRLINE FARE PRICE PREDICTION
AIRLINE FARE PRICE PREDICTIONAIRLINE FARE PRICE PREDICTION
AIRLINE FARE PRICE PREDICTIONIRJET Journal
 
19004 Practice school finalReportformat Model (1).pdf
19004 Practice school finalReportformat Model (1).pdf19004 Practice school finalReportformat Model (1).pdf
19004 Practice school finalReportformat Model (1).pdfAkshayKumar983983
 
Traffic Sign Recognition Model
Traffic Sign Recognition ModelTraffic Sign Recognition Model
Traffic Sign Recognition ModelIRJET Journal
 
Traffic Flow Prediction Using Machine Learning Algorithms
Traffic Flow Prediction Using Machine Learning AlgorithmsTraffic Flow Prediction Using Machine Learning Algorithms
Traffic Flow Prediction Using Machine Learning AlgorithmsIRJET Journal
 
IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...
IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...
IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...IRJET Journal
 
A Comparative Study on Identical Face Classification using Machine Learning
A Comparative Study on Identical Face Classification using Machine LearningA Comparative Study on Identical Face Classification using Machine Learning
A Comparative Study on Identical Face Classification using Machine LearningIRJET Journal
 
IRJET - Obstacle Detection using a Stereo Vision of a Car
IRJET -  	  Obstacle Detection using a Stereo Vision of a CarIRJET -  	  Obstacle Detection using a Stereo Vision of a Car
IRJET - Obstacle Detection using a Stereo Vision of a CarIRJET Journal
 
IRJET- Parking Space Detection using Image Processing in MATLAB
IRJET- Parking Space Detection using Image Processing in MATLABIRJET- Parking Space Detection using Image Processing in MATLAB
IRJET- Parking Space Detection using Image Processing in MATLABIRJET Journal
 
Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...
Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...
Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...MIbrar4
 
IRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural NetworksIRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural NetworksIRJET Journal
 
Obstacle Detection and Collision Avoidance System
Obstacle Detection and Collision Avoidance SystemObstacle Detection and Collision Avoidance System
Obstacle Detection and Collision Avoidance SystemIRJET Journal
 
deep-reinforcement-learning-framework.pdf
deep-reinforcement-learning-framework.pdfdeep-reinforcement-learning-framework.pdf
deep-reinforcement-learning-framework.pdfYugank Aman
 
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...IRJET Journal
 
IRJET- American Sign Language Classification
IRJET- American Sign Language ClassificationIRJET- American Sign Language Classification
IRJET- American Sign Language ClassificationIRJET Journal
 

Similar to IRJET-Survey on Simulation of Self-Driving Cars using Supervised and Reinforcement Learning Models (20)

IRJET- Self-Driving Cars: Automation Testing using Udacity Simulator
IRJET- Self-Driving Cars: Automation Testing using Udacity SimulatorIRJET- Self-Driving Cars: Automation Testing using Udacity Simulator
IRJET- Self-Driving Cars: Automation Testing using Udacity Simulator
 
IRJET - Steering Wheel Angle Prediction for Self-Driving Cars
IRJET - Steering Wheel Angle Prediction for Self-Driving CarsIRJET - Steering Wheel Angle Prediction for Self-Driving Cars
IRJET - Steering Wheel Angle Prediction for Self-Driving Cars
 
Self Driving Car
Self Driving CarSelf Driving Car
Self Driving Car
 
Self-Driving Car to Drive Autonomously using Image Processing and Deep Learning
Self-Driving Car to Drive Autonomously using Image Processing and Deep LearningSelf-Driving Car to Drive Autonomously using Image Processing and Deep Learning
Self-Driving Car to Drive Autonomously using Image Processing and Deep Learning
 
Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...
Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...
Comparative Analysis of Tuning Hyperparameters in Policy-Based DRL Algorithm ...
 
AIRLINE FARE PRICE PREDICTION
AIRLINE FARE PRICE PREDICTIONAIRLINE FARE PRICE PREDICTION
AIRLINE FARE PRICE PREDICTION
 
19004 Practice school finalReportformat Model (1).pdf
19004 Practice school finalReportformat Model (1).pdf19004 Practice school finalReportformat Model (1).pdf
19004 Practice school finalReportformat Model (1).pdf
 
Traffic Sign Recognition Model
Traffic Sign Recognition ModelTraffic Sign Recognition Model
Traffic Sign Recognition Model
 
Traffic Flow Prediction Using Machine Learning Algorithms
Traffic Flow Prediction Using Machine Learning AlgorithmsTraffic Flow Prediction Using Machine Learning Algorithms
Traffic Flow Prediction Using Machine Learning Algorithms
 
IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...
IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...
IRJET- A Review on Deep Reinforcement Learning Induced Autonomous Driving Fra...
 
A Comparative Study on Identical Face Classification using Machine Learning
A Comparative Study on Identical Face Classification using Machine LearningA Comparative Study on Identical Face Classification using Machine Learning
A Comparative Study on Identical Face Classification using Machine Learning
 
IRJET - Obstacle Detection using a Stereo Vision of a Car
IRJET -  	  Obstacle Detection using a Stereo Vision of a CarIRJET -  	  Obstacle Detection using a Stereo Vision of a Car
IRJET - Obstacle Detection using a Stereo Vision of a Car
 
IRJET- Parking Space Detection using Image Processing in MATLAB
IRJET- Parking Space Detection using Image Processing in MATLABIRJET- Parking Space Detection using Image Processing in MATLAB
IRJET- Parking Space Detection using Image Processing in MATLAB
 
Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...
Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...
Demonstrate the implementation PI controller to regulate speed of DC Servo Mo...
 
IRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural NetworksIRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural Networks
 
Obstacle Detection and Collision Avoidance System
Obstacle Detection and Collision Avoidance SystemObstacle Detection and Collision Avoidance System
Obstacle Detection and Collision Avoidance System
 
Mini Project- Stepper Motor Control
Mini Project- Stepper Motor ControlMini Project- Stepper Motor Control
Mini Project- Stepper Motor Control
 
deep-reinforcement-learning-framework.pdf
deep-reinforcement-learning-framework.pdfdeep-reinforcement-learning-framework.pdf
deep-reinforcement-learning-framework.pdf
 
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...
IRJET- Prediction of Crime Rate Analysis using Supervised Classification Mach...
 
IRJET- American Sign Language Classification
IRJET- American Sign Language ClassificationIRJET- American Sign Language Classification
IRJET- American Sign Language Classification
 

More from IRJET Journal

TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...IRJET Journal
 
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURESTUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTUREIRJET Journal
 
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...IRJET Journal
 
Effect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsEffect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsIRJET Journal
 
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...IRJET Journal
 
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...IRJET Journal
 
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...IRJET Journal
 
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...IRJET Journal
 
A REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASA REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASIRJET Journal
 
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...IRJET Journal
 
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProP.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProIRJET Journal
 
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...IRJET Journal
 
Survey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemSurvey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemIRJET Journal
 
Review on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesReview on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesIRJET Journal
 
React based fullstack edtech web application
React based fullstack edtech web applicationReact based fullstack edtech web application
React based fullstack edtech web applicationIRJET Journal
 
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...IRJET Journal
 
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.IRJET Journal
 
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...IRJET Journal
 
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignMultistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignIRJET Journal
 
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...IRJET Journal
 

More from IRJET Journal (20)

TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
TUNNELING IN HIMALAYAS WITH NATM METHOD: A SPECIAL REFERENCES TO SUNGAL TUNNE...
 
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURESTUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
STUDY THE EFFECT OF RESPONSE REDUCTION FACTOR ON RC FRAMED STRUCTURE
 
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
A COMPARATIVE ANALYSIS OF RCC ELEMENT OF SLAB WITH STARK STEEL (HYSD STEEL) A...
 
Effect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil CharacteristicsEffect of Camber and Angles of Attack on Airfoil Characteristics
Effect of Camber and Angles of Attack on Airfoil Characteristics
 
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
A Review on the Progress and Challenges of Aluminum-Based Metal Matrix Compos...
 
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
Dynamic Urban Transit Optimization: A Graph Neural Network Approach for Real-...
 
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
Structural Analysis and Design of Multi-Storey Symmetric and Asymmetric Shape...
 
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
A Review of “Seismic Response of RC Structures Having Plan and Vertical Irreg...
 
A REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADASA REVIEW ON MACHINE LEARNING IN ADAS
A REVIEW ON MACHINE LEARNING IN ADAS
 
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
Long Term Trend Analysis of Precipitation and Temperature for Asosa district,...
 
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD ProP.E.B. Framed Structure Design and Analysis Using STAAD Pro
P.E.B. Framed Structure Design and Analysis Using STAAD Pro
 
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
A Review on Innovative Fiber Integration for Enhanced Reinforcement of Concre...
 
Survey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare SystemSurvey Paper on Cloud-Based Secured Healthcare System
Survey Paper on Cloud-Based Secured Healthcare System
 
Review on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridgesReview on studies and research on widening of existing concrete bridges
Review on studies and research on widening of existing concrete bridges
 
React based fullstack edtech web application
React based fullstack edtech web applicationReact based fullstack edtech web application
React based fullstack edtech web application
 
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
A Comprehensive Review of Integrating IoT and Blockchain Technologies in the ...
 
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
A REVIEW ON THE PERFORMANCE OF COCONUT FIBRE REINFORCED CONCRETE.
 
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
Optimizing Business Management Process Workflows: The Dynamic Influence of Mi...
 
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic DesignMultistoried and Multi Bay Steel Building Frame by using Seismic Design
Multistoried and Multi Bay Steel Building Frame by using Seismic Design
 
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
Cost Optimization of Construction Using Plastic Waste as a Sustainable Constr...
 

Recently uploaded

Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort servicejennyeacort
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfAsst.prof M.Gokilavani
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝soniya singh
 
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escortsranjana rawat
 
power system scada applications and uses
power system scada applications and usespower system scada applications and uses
power system scada applications and usesDevarapalliHaritha
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024Mark Billinghurst
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
 
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2RajaP95
 
Current Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLCurrent Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLDeelipZope
 
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube ExchangerStudy on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube ExchangerAnamika Sarkar
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...srsj9000
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AIabhishek36461
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Dr.Costas Sachpazis
 
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130Suhani Kapoor
 
Introduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxIntroduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxvipinkmenon1
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxwendy cai
 

Recently uploaded (20)

Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort serviceGurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
Gurgaon ✡️9711147426✨Call In girls Gurgaon Sector 51 escort service
 
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdfCCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
CCS355 Neural Network & Deep Learning Unit II Notes with Question bank .pdf
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
 
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
 
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
 
power system scada applications and uses
power system scada applications and usespower system scada applications and uses
power system scada applications and uses
 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024
 
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Meera Call 7001035870 Meet With Nagpur Escorts
 
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2HARMONY IN THE HUMAN BEING - Unit-II UHV-2
HARMONY IN THE HUMAN BEING - Unit-II UHV-2
 
Current Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCLCurrent Transformer Drawing and GTP for MSETCL
Current Transformer Drawing and GTP for MSETCL
 
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube ExchangerStudy on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
Study on Air-Water & Water-Water Heat Exchange in a Finned Tube Exchanger
 
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
Gfe Mayur Vihar Call Girls Service WhatsApp -> 9999965857 Available 24x7 ^ De...
 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AI
 
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
Sheet Pile Wall Design and Construction: A Practical Guide for Civil Engineer...
 
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
VIP Call Girls Service Hitech City Hyderabad Call +91-8250192130
 
Introduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptxIntroduction to Microprocesso programming and interfacing.pptx
Introduction to Microprocesso programming and interfacing.pptx
 
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCRCall Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
Call Us -/9953056974- Call Girls In Vikaspuri-/- Delhi NCR
 
What are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptxWhat are the advantages and disadvantages of membrane structures.pptx
What are the advantages and disadvantages of membrane structures.pptx
 
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Serviceyoung call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
young call girls in Rajiv Chowk🔝 9953056974 🔝 Delhi escort Service
 
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
🔝9953056974🔝!!-YOUNG call girls in Rajendra Nagar Escort rvice Shot 2000 nigh...
 

IRJET-Survey on Simulation of Self-Driving Cars using Supervised and Reinforcement Learning Models

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2633 Survey on Simulation of Self-Driving Cars using Supervised and Reinforcement Learning Models Athul Dev1, Roshni K S2 1 Student, Department of Computer Science & Engineering, Nitte Meenakshi Institute of Technology, India 2 Student, Dept. of Electrical & Electronics Engineering, Sir. M Vishveshvaraya Institute of Technology, India -----------------------------------------------------------------------------***--------------------------------------------------------------------------- Abstract - A self-driving system is the heart, soul and futureof the automobile industry and there is a lot of techniques and technology involved in the accomplishment of a self-driving car. Since the system has to be deployed in the real world and which is supposed to take multiple actions and sudden reactions in real time, the system should be tuned to a very high degree of precision. In this paper we share our observations with respect to tuning and tweaking most of the hyper parameters and the architecture in a generated environment for both supervised and reinforcement learning techniques. And how different a supervised learning model is to a reinforcement learning model with respect to the self- driving car. We use the concepts of Deep Learning, Deep Neural Networks, Convolutional Networks and Deep Q learning in order to configure the brain of our car to perform certain actions. Key Words: Machine Learning, Supervised Learning, Reinforcement Learning, Q Learning, Convolutional Neural Networks, Deep Learning, Markov Decision Process. 1.INTRODUCTION From the perspective of a self-driving car, we have mainly considered the motion, the movement and the overall physics of the car. In the supervised model wehavemadethe prediction of the car’s position, its steering angle and the speed of the car in the environment based on how we drove the car in that particular environment, so on the basis of certain frames of images encountered, the required steering angle and speed in fed back to the car as it’s input. That is image is the input and the steering angle with the speed, throttle and brake is the output. So in this case we are not considering any sorts of external obstacles like another vehicle, a barricade and so on. Whereasin the reinforcement model we have considered the obstacles by giving it a negative reward when it encounters and obstacles or barriers, due to which it tends to avoid these obstacles. We have tuned various hyper parameters like the learning rate, living penalty, rewards, the activation functions and the network architecture. And the testing for both is done for various time intervals with its related accuracy value. The combination of the environments or the learning techniquescan be made possible which is the futurescopeof this paper. 2. SUPERVISED LEARNING MODEL Supervised Learning is the machine learning taskoflearning a function that mapsan input to an output based on example input-output pairs. It infers a function from labeled training data consisting of a set of training examples. In this case the imagesrecorded from the car when riddenis the input data and the steering angle is the target label. And this supervised model in general is a technique that was implemented by the [1]NVIDIA’sEnd-to-End Deep Learning for Self-Driving Cars. We have used the concept and modelled a system with different activation functions and different convolutional layers. 2.1 Implementation The implementation is done in three main stages/phases: a. Data Generation Phase b. Training Phase c. Testing Phase 2.1.1 Data Generation Phase For the data generation part, where the data collectedhereis the imagesfrom the camera(s), which is nothingbuttheroad that the car sees to its front and various physics variables likeitsspeed, throttle, brake and steering. In order to getthe data and for the environment we have used Udacity’s open source version of the car simulator which comes with two pre-loaded environments and an efficient way to train and test the model. In order to generate the data we select the training mode of the open source simulator which brings up a car with three camera’s mounted on it, towards the left, right and center of the car. So we start moving the car with the keyboard inputs
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2634 or with a joystick for smother transitions,wecannowrecord this movement of the car which is controlled by us(ahuman driver) and this action will be mimicked by ourmodel in the form of mapping images to actions. This data can then be saved into a comma separated value file, which has the images from the camera’s pointing to left, right and center of the car with the image’s absolute path, the speed, throttle, brake and the steering values 2.1.2 Training Phase The training process is inspired by the [1] NVIDIA’s End-to- End Deep Learning for Self-Driving Cars, they have made use of an actual car fitted with three camerason itsleft, right and center and had a human driver drive the car for 72hrs, recording the images from the cameras along with the car’s speed, throttle, brake and the steering angle values. We have mimicked the above process using the Udacity’s open source version of the car simulator, by making a user (human driver) drive the car in the simulator using a keyboardfor various lapsand different intervalsoftime.And have recorded it with respect to the number of laps and various time intervals which is shown in the Table 1. Table -1: Accuracy with respect to the no. of laps. Number of Laps Time taken to train the model Accuracy of the model 1 15mins 20% 3 50mins 45% 5 3hrs 70% 10 8hrs 85% In the model, we have used Keras for our deep learning framework with a sequential model. We start with doing image normalization and adding a 10 layer convolutional network layer with 5 fully connected layers to flatten the data. We have tried the model using various activation functions like the sigmoid activation function, rectifiedlinear activation function and the exponential linearunit activation function, and found exponential linearunitthebestactivation function for this scenario as it takes care of the vanishing gradient problem. Table -2: Accuracy with respect to the activation functions Activation Function Accuracy of the model Sigmoid 70-85% Rectified Linear Unit 80-90% Exponential Linear Unit 85-95% Now since the setup is ready, we try to find the data loss or the error that is by finding the difference between the predicted steering command and the actual steering command from the training data. Once we determine the mean of sumof squared error,that isthenbackpropagatedto the convolutional neural network which in turn produces a network computed steering value. And for optimization we have used Adam optimizer that does our Gradient decent function. 2.1.3 Testing Phase A server-client structure is used for the testing, where the server is the open source simulator and the client, our python program. Variousdependencies and librarieslikesocketiowhichhelps us work in terms of piping commands in real time using event handlers, Pillow for image manipulation and flask which is a web framework. For testing, we start our simulator and then pass the model that was trained in the training phase to our testing script and simultaneously activate the autonomous mode in the simulator to which the program connects and outputs the required steering angle, speed, throttle and brake values to the autonomous car. Here the prediction of the steering angle with respect to the image from the camerasis sent to the server and Thusgiving us a self-driving car in a simulated environment. Fig -1: Testing the model
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2635 3. REINFORCEMENT LEARNING MODEL Reinforcement learning is a type of Machine Learning algorithms which allows software agents and machines to automatically determine the ideal behavior within a specific context, to maximize its performance. Reinforcement algorithms are not given explicit goals; instead, they are forced to learn these optimal goals by trial and error. Here we are makinguse ofthe sole conceptsofreinforcement learning like theBellman Equation,MarkovDecisionProcess, and Q Learning which is the base for Deep Q Learning Networks. 3.1 Concepts 3.1.1 Bellman Equation Say we have a state s, an action taken by the agent as a, the reward R which the agent gets once it gets into a particular state and Ὑ (gamma) the discount. The bellman equation assigns a value to a statebygettingthe maximum of all the actions possible with respect to the sum of the rewards that the agent gets in the current state for a certain action and the value of the following stateorthestate that the agent would end up in from the current state. 3.1.2 Markov Decision Process The Markov Decision Process takes care of the non- deterministic search problem. That is the agent is bound the take any action in order to find itsoptimal policy or pathand the actions it might take may not be in a deterministic way, thus we have a probability term that is multiplied to the value of the following state. 3.1.3 Living Penalty Generally an agent gets a reward say in our case a positive reward when it reaches its destinations and a negative reward when it goes off road, but a living penalty is a reward that is given to an agent throughout in its exploration phase so as to generally find an optimal path. The policy or the mapping changes with respect to the living penalty value that is assigned to the agent in its environment. 3.1.4 Q Learning Q-learning is a simple way for agents to learn how to act optimally in controlled Markovian domains.Itamountstoan incremental method for dynamic programming which imposes limited computational demands. It works by successively improving its evaluations of the quality of particular actions at particular states. The value of the states are replaced by the Q functions or rather the Quality factor of the action in the respective states. 3.2 Design and Environment We have made use of Kivy which is an open source Python library for development of our self-driving car interface or the environment and for the mapping purposes. In the environment we have a car that goes from one destination to another, where in this casethedestinationsare predefined. Asit movesitis encounterstomanybarriersand the car overcomes or surpasses these barriers and continue its route using the idea of Deep Q learning or in general Reinforcement learning technique(s). The user is allowed to place or insert the barriers in the course of the car and the car surpasses these barriers in real time with minimal fault and training. The learning, with respect to this scenario is very fast as compared to the previously mentioned supervised learning model. The possible actions that the car can take here is move left, right or straight. And the angle of the rotation to its left or right is specified as 10. The car is equipped with three sensors which are to its left, right and center and these sensors detect the barriers and avoids them by taking the opposite direction of the sensor that detects the barriers. We also configure the rewards that the car gets when it hits the barrier as-1whichis theworstreward, the livingpenalty reward as -0.4, if it reaches its optimal path that is a path better than the previousone in termsof the distance taken it gets a reward of 0.2, a reward of -1 if it hits the environment window or the limit and a reward of 1 if it reaches its destination. 3.3 Implementation In the reinforcement learning there is no need of explicit data generation because the data is obtained as the agent is exploring, and the training and testing happens simultaneously.
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2636 We are extensively using PyTorch deep learning framework for creating the neural network architecture. The neural network architecture is comprising of 2 fully connected layers in a linear manner with 30 nodes in the hidden layer, the first fully connected layer is with input and the hidden layer and the next is between the hidden layer and the number of actions the car can take. Next we start building the feed forward network, for this we apply an activation function toour first fully connectedlayer and generate the Q values by passing the output of the activation function to the second fully connected layer. And return the Q values. This architecture can be improved by tweaking the number of nodesin the hiddenlayersandusing different activation functions. Then we continue to implement an experience reply for the agent as the Markov Decision process or the Q learning algorithm works with respect to analyzing a seriesof events and the consecutive states, that is the current state and the following state. The correlation among these statesaremore and the network will not be learning very well. So inorderto avoid this we use experience reply which considers say about 100-1000 statesand keepsthem in the memoryasthe series of events so that an efficient decision can be taken by the agent and these bulk states are treated in various batches. Finally we build our Deep Q learning model by passing various hyper parameterslikethe discount factor (Gamma), the model itself from the neural network architecture, the reply memory and the optimizer to do stochastic gradient decent where in this case an Adam Optimizer is used with the learning rate of 0.005 which gave the best result.Wealso determine the right action for the car to take that is whether to go left, right or straight using a softmax layer. Softmax gives a large probability to highest Q value. Softmax also takes care of the exploration of the environmentbytheagent really well. The learning is done in batches with respect to the reply memory technique. While learning weconsiderthe maximum of the Q values of the next state as well as the reward obtained in the current state. We then calculate the loss using the smooth l1 loss (Huber loss - since it improves the Q learning) and pass the loss error to the optimizer todo the stochastic gradient decent and update the weights using the back propagation. Eventually an update is done with respect to the action and the state when the agent gets intoa new state and a connection is made between the brain (the DQN logic) and the map function which is defined earlier in the section 3.2 Design and Environment. 3.4 Training and Testing The training and testing can be visualized once we run the environment and the car, since the Deep Q Network (DNQ) logic is linked to the mapping part we need not explicitly pass any arguments to link them at the time of execution. The car starts exploring the environment for some time and once it reachesitsdestination it gets a positiverewardandit looksfor itsnext destination checkpoint andcontinuestoget positive reward from reaching checkpoints and finds an optimal path to do the same. But when we start placing barriers in the environment and when the car encounters these barriers it receives a negative reward, so in its next encounter it evades the barrier and becomes much efficient to its previous transition. The best part about this system is that the barriers or the obstacles can be randomly anytime and anywhere on the environment and the car or the agent avoidsor evadesthese barriers in real time. Fig -2: Bird’s view of the environment Thus, the agent updates the rewards and the scores continuously andtakesthe next actions with respect tothese updated rewards, scores and the Q values. This gives rise to a self-driving car using the reinforcement learning technique. Chart -1: Behavior with respect to the time and the reward
  • 5. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 06 | June 2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 7.211 | ISO 9001:2008 Certified Journal | Page 2637 4. SUPERVISED V/S REINFORCEMENT MODEL In the supervised learning model the mapping between the images to the car’s attributes like speed, steering angle, throttle and brake is done, so whenever the car/ model gets a new image it will act accordingly whereasinreinforcement learning model the car explores the environment at first for some time as a part of training and with respect to the rewards it gets, it moves in the environment taking various decisions like going left, right or straight. Supervised learning model needs a lot of data and predefined training for a long time only then it can be implement accurately whereas reinforcement learning adapts quickly with less training and data. Reinforcement learning model in this scenario can easily evade or surpass any obstacle thrown to it by the user anytime, anywhere whereas the supervised learning model fails to do so. Supervised learning model can easily facilitate a large number of features and a high accuracy can be obtained but we will have to use a powerful enough computerthat run on multiple graphics processing unit(s). 5. CONCLUSIONS In this paper we have given an overview of the two different models that is the supervised learning model and the reinforcement learning model with respect to a self-driving car. And how each model behaves when the hyper parameters are tweaked, variations in the activation functions is made and how long the model is trained. The model can be furthered improvised by changing the environments and the architecture, and the implementation in general can be improved by combining the learning models and further tweaking the hyper parameters or adding more layers to the architecture. ACKNOWLEDGEMENT We thank our family and friends for their priceless support and encouragement. REFERENCES [1] NVIDIA’s End-to-End Deep Learning for Self-Driving Cars. [2] The Theory of Dynamic Programming By Richard Bellman (1954) [3] Markov Decision Processes:ConceptsandAlgorithmsBy Martijn van Otterlo (2009) [4] Learning to Predict by the Methods of Temporal Differences BY Richard Sutton (1988) [5] Reinforcement Learning I: Introduction By Richard Sutton. [6] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. R. Howard, W. Hubbard and L. D. Jackel. Backprop-agation applied to handwritten zip code recognition. BIOGRAPHIES Athul Dev Machine Learning and Data Science enthusiast. Experiences: Software Developer Intern at Power Grid Corporation of India, Machine Learning Intern at Latent Talent Technologies and Curl Analytics, Data Analyst Intern at CHR Solutions. Interested in Big Data Analytics, AI, ML and web development. Roshni K S IoT, embedded systems, microcontrollers,matlabandcomputer programming enthusiast. Experiences: Software Developer Intern at Power Grid Corporation of India, Power System Analyst Intern at KPTCL.