SlideShare a Scribd company logo
An Efficient use of temporal difference technique in computer game
learning
Indian institute of technology
( Indian school of mines ),
Dhanbad.
Project guide:- Presented by:
Dr. Rajendra Pamula Prabhu Kumar
Department of computer 15MT000624
Science and engineering Computer science and engineering
Indian institute of technology Indian institute of technology
( Indian school of mines), ( Indian school of mines),
Dhanbad Dhanbad
Outline
1. Introduction of reinforcement learning
2. Agent-Environment interface
3. Types of reinforcement learning
4. Elements of the reinforcement learning
5. Types of selection of state
6. Algorithms of reinforcement learning
References
Introduction of reinforcement learning
 Reinforcement learning is the part of machine learning, which is a field of computer
science that gives computer to ability to learn without being explicitly programmed.
 Reinforcement learning is a framework for computational learning agents use experience
from their interaction with an environment to improve performance over time.
 In reinforcement learning task, the agent understands the state of the environment and it
always tries to maximize the long-term return which is based on real value reward.
 It is learning of what to do-how to do mapping situation to action so as maximize total
numerical reward and minimize the penalty.
Introduction of reinforcement learning cont.…
• If there is no explicit teacher to guide the learning agent, the agent must learn the behavior
through trail-and-error interaction with unknown environment.
• The learning agent senses the environment, takes actions on it, and receives numeric reward or
punishment from some reward function.
• When we say agent learn means ”sometimes it modifies the code itself or modifies the database ”,
database implies the experiences, information, event etc.
• It is responsible for making decision.
• The main goal of reinforcement learning is “Buildup a good model such as algorithm which
generate a sequence of decision and lead to the highest long-term reward.”
Agent environment interface
o At each time step t, the reinforcement learning agent
receives some representation of environment’s current state
s(t) € S ,where S is the set of possible state and then choose
some action a(t)€ A(st), where A(st) is set of actions that can
be executed in state s(t).
o The agent receives reward r(t+1) and execute in next state
s(t+1)
o The reward function can be used for specify the wide range
of planning goals, It means the designer can tell the agent
what he has achieve.
o The reward function which must be unalterable by the
agent.
Types of reinforcement learning
There are two types of reinforcement learning
1. Episodic: The interaction with the environment is divided into independent episodes.
“Independent”, means performance in each episode is depends only the action taken on that
episode.
in episodic task, a return is sum of all reward received from the beginning of the episode until ends.
where, T is terminal state i.e. end of episode ends
S0 is the starting state of
R denotes as total return
r(k) denotes as the reward on the kth states
Types of the reinforcement learning contd..
2.Continuing task: It consist infinite sequence of state, action and rewards. In this task, the action
and environment interaction doesn’t break down in separate episode. The performance is depends
upon the current action.
In the case of continuing task, The return is depends upon discount factor
 where γ denotes discount factor which adjust the relative importance
between long-term consequences vs. short term consequences.
 The discount factor is between 0 and 1.
 The discount factor reflects the strategy of how fast learning takes place
 If γ =0, agent only concerned about maximizing the immediate rewards
 If γ approaches to 1, The agent takes the future reward into account
Element of the reinforcement learning
1. Policy:
It defines the learning agent’s way of behaving at a given time.
It might be a function or simple lookup table.
It only used in reinforcement learning is to determine the behavior.
2. Reward function:
It is the function which defines which one is the bad and good event for agent.
It maps each state-action pair of the environment to a single real number.
It must necessarily unalterable by the agent.
Elements in the reinforcement learning contd..
3. Value function:
It specifies what is good in long run. The value of the state is the total amount of reward an
agent can expect to accumulate over the future, starting from that state.
Where as reward is the immediate desirability of environmental states. i.e. values indicate the
long-tem desirability of states.
4. Model:
It is used for planning. It defines the copy of behavior, e.g. by given state and action ,The model
might predict the next state and next reward.
Algorithm of reinforcement learning
1. Markov decision processes:
• It is standard, general formalism for sequential decision problems.
• It consist tuple of <S,A,P,R>
where S is the set of states.
A is the set of actions available to the agent
P is the probability, P(a, ss′) = P r {st+1 = s ′ | st = s, at = a}, it is a state transition function that
defines the probability of transitioning to state s ′ at time t + 1 after action a is taken when agent is
in state s at time t.
R is the reward function that determines the probability of receiving reward after choosing action a
in state s and going for next state s’.
Algorithms in reinforcement learning
2. Dynamic programming (DP)
• It is the method to solve the markov decision process i.e. to find an optimal policy, if the full
knowledge of model is available.
• For dynamic programming, all the transition probabilities and reward expectation must be known.
• This algorithm updates the estimates of states values based on their estimates of the next state.
• There are two basic DP methods used for computing optimal policy
1. Policy iteration
2. Value iteration
Policy Iteration:
• It forms a sequence of policy Ωo, Ω1, Ω2….Ωk, Ωk+1 where Ωk+1 is an improvement of Ωk.
• Policy evaluation task is concerned with computing state value function for any policy Ω
• The iterative algorithm for policy evaluation is
• Estimating value functions is particularly useful for finding the better policy.
• The policy improvement algorithm uses action-value function to improve the current policy. If
then it is better to select
action a in policy Ω
• If Ω and Ω’ are two policy and this condition hold then Ω’ is the better policy than Ω
Value iteration
• In value iteration, optimal policy is not computed directly.
• For that, optimal value function is computed and then a greedy policy with respect to function is
an optimal policy.
• It stops and find the optimal policy when the changes introduced by backups/updates becomes
sufficiently small.
• One threshold value has been initialized and compared with threshold value.
• If the policy value is sufficiently smaller than threshold, the policy is called as optimal policy
3. Temporal difference
• The temporal differences idea has been taken from dynamic programming.
• The temporal difference and dynamic programming, both are used for accumulating the value functions.
• In this methods, learning takes place after every time step which is beneficial as it makes for efficient
learning
• The agent can revise its policy after every action and state it experiences.
• TD algorithms make updates of the estimated policy values based on each state transition and on the
immediate reward received from the environment on this transition.
• The initial temporal difference algorithm is called TD(0),called tabular estimates v(Ω).
It updates by following method
V (s) ← V (s) + α (V (s’) − V (s)) where α is the positive step size parameter
V(s’) is the value function for next state
α (V (s’) − V (s)), called as Temporal difference error. V(s) is the value function for the current state
Which is always designed to move toward 0.
Type of selection of states
1. Greedy
2. Exploration process
a) Providing initial knowledge
b) Deriving a policy from demonstration
c) Ask for help
d) Teacher provide advice
Application of reinforcement learning
1. Benchmark Problems
a) Mountain car
b) Cart-pole balancing
c) Pendulum swing up tec..
2. Games
a) Tic-Tac-Toe
b) Chess etc..
3. Real world applications
a) Robotics
b) Control of helicopter
c) Prediction of stock prices
References
• R. S. Sutton. Reinforcement learning: past, present and future [online]. Available from http:
//www-anw. cs. umass. edu/ rich/Talks/SEAL98/SEAL98. html [accessed on December 2005]. 1999.
• R. S. Sutton and A. G. Barto. Reinforcement learning. an introduction. Cambridge, MA: The MIT
Press, 1998.
• M. L. Puterman. Markov decision processes-discrete stochastic dynamic programming. John
Wiley and sons, Inc, New York, NY, 1994.
Thank You !!

More Related Content

What's hot

Reinforcement Learning 8: Planning and Learning with Tabular Methods
Reinforcement Learning 8: Planning and Learning with Tabular MethodsReinforcement Learning 8: Planning and Learning with Tabular Methods
Reinforcement Learning 8: Planning and Learning with Tabular Methods
Seung Jae Lee
 
Introduction: Asynchronous Methods for Deep Reinforcement Learning
Introduction: Asynchronous Methods for  Deep Reinforcement LearningIntroduction: Asynchronous Methods for  Deep Reinforcement Learning
Introduction: Asynchronous Methods for Deep Reinforcement Learning
Takashi Nagata
 
Optimization Using Evolutionary Computing Techniques
Optimization Using Evolutionary Computing Techniques Optimization Using Evolutionary Computing Techniques
Optimization Using Evolutionary Computing Techniques
Siksha 'O' Anusandhan (Deemed to be University )
 
Proximal Policy Optimization Algorithms, Schulman et al, 2017
Proximal Policy Optimization Algorithms, Schulman et al, 2017Proximal Policy Optimization Algorithms, Schulman et al, 2017
Proximal Policy Optimization Algorithms, Schulman et al, 2017
Chris Ohk
 
A brief introduction to Searn Algorithm
A brief introduction to Searn AlgorithmA brief introduction to Searn Algorithm
A brief introduction to Searn Algorithm
Supun Abeysinghe
 
ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...
ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...
ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...
Ajay Kumar
 
Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021
Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021
Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021
Chris Ohk
 
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
Chris Ohk
 
Reinforcement Learning 5. Monte Carlo Methods
Reinforcement Learning 5. Monte Carlo MethodsReinforcement Learning 5. Monte Carlo Methods
Reinforcement Learning 5. Monte Carlo Methods
Seung Jae Lee
 
Metaheuristics
MetaheuristicsMetaheuristics
Metaheuristics
ossein jain
 
Lecture notes
Lecture notesLecture notes
Lecture notesbutest
 
Financial Trading as a Game: A Deep Reinforcement Learning Approach
Financial Trading as a Game: A Deep Reinforcement Learning ApproachFinancial Trading as a Game: A Deep Reinforcement Learning Approach
Financial Trading as a Game: A Deep Reinforcement Learning Approach
謙益 黃
 
Introduction to dynamic programming
Introduction to dynamic programmingIntroduction to dynamic programming
Introduction to dynamic programming
Amisha Narsingani
 
Multiobjective optimization and Genetic algorithms in Scilab
Multiobjective optimization and Genetic algorithms in ScilabMultiobjective optimization and Genetic algorithms in Scilab
Multiobjective optimization and Genetic algorithms in Scilab
Scilab
 
Linear Programming Problems {Operation Research}
Linear Programming Problems {Operation Research}Linear Programming Problems {Operation Research}
Linear Programming Problems {Operation Research}
FellowBuddy.com
 
Multiobjective optimization and trade offs using pareto optimality
Multiobjective optimization and trade offs using pareto optimalityMultiobjective optimization and trade offs using pareto optimality
Multiobjective optimization and trade offs using pareto optimality
Amogh Mundhekar
 
Multi Objective Optimization
Multi Objective OptimizationMulti Objective Optimization
Multi Objective Optimization
Nawroz University
 
Multiobjective presentation
Multiobjective presentationMultiobjective presentation
Multiobjective presentation
Mohammed Kamil
 
001 lpp introduction
001 lpp introduction001 lpp introduction
001 lpp introduction
Victor Seelan
 

What's hot (20)

Reinforcement Learning 8: Planning and Learning with Tabular Methods
Reinforcement Learning 8: Planning and Learning with Tabular MethodsReinforcement Learning 8: Planning and Learning with Tabular Methods
Reinforcement Learning 8: Planning and Learning with Tabular Methods
 
Introduction: Asynchronous Methods for Deep Reinforcement Learning
Introduction: Asynchronous Methods for  Deep Reinforcement LearningIntroduction: Asynchronous Methods for  Deep Reinforcement Learning
Introduction: Asynchronous Methods for Deep Reinforcement Learning
 
Optimization Using Evolutionary Computing Techniques
Optimization Using Evolutionary Computing Techniques Optimization Using Evolutionary Computing Techniques
Optimization Using Evolutionary Computing Techniques
 
Proximal Policy Optimization Algorithms, Schulman et al, 2017
Proximal Policy Optimization Algorithms, Schulman et al, 2017Proximal Policy Optimization Algorithms, Schulman et al, 2017
Proximal Policy Optimization Algorithms, Schulman et al, 2017
 
A brief introduction to Searn Algorithm
A brief introduction to Searn AlgorithmA brief introduction to Searn Algorithm
A brief introduction to Searn Algorithm
 
ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...
ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...
ADVANCED OPTIMIZATION TECHNIQUES META-HEURISTIC ALGORITHMS FOR ENGINEERING AP...
 
Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021
Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021
Adversarially Guided Actor-Critic, Y. Flet-Berliac et al, 2021
 
CS799_FinalReport
CS799_FinalReportCS799_FinalReport
CS799_FinalReport
 
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
Continuous Control with Deep Reinforcement Learning, lillicrap et al, 2015
 
Reinforcement Learning 5. Monte Carlo Methods
Reinforcement Learning 5. Monte Carlo MethodsReinforcement Learning 5. Monte Carlo Methods
Reinforcement Learning 5. Monte Carlo Methods
 
Metaheuristics
MetaheuristicsMetaheuristics
Metaheuristics
 
Lecture notes
Lecture notesLecture notes
Lecture notes
 
Financial Trading as a Game: A Deep Reinforcement Learning Approach
Financial Trading as a Game: A Deep Reinforcement Learning ApproachFinancial Trading as a Game: A Deep Reinforcement Learning Approach
Financial Trading as a Game: A Deep Reinforcement Learning Approach
 
Introduction to dynamic programming
Introduction to dynamic programmingIntroduction to dynamic programming
Introduction to dynamic programming
 
Multiobjective optimization and Genetic algorithms in Scilab
Multiobjective optimization and Genetic algorithms in ScilabMultiobjective optimization and Genetic algorithms in Scilab
Multiobjective optimization and Genetic algorithms in Scilab
 
Linear Programming Problems {Operation Research}
Linear Programming Problems {Operation Research}Linear Programming Problems {Operation Research}
Linear Programming Problems {Operation Research}
 
Multiobjective optimization and trade offs using pareto optimality
Multiobjective optimization and trade offs using pareto optimalityMultiobjective optimization and trade offs using pareto optimality
Multiobjective optimization and trade offs using pareto optimality
 
Multi Objective Optimization
Multi Objective OptimizationMulti Objective Optimization
Multi Objective Optimization
 
Multiobjective presentation
Multiobjective presentationMultiobjective presentation
Multiobjective presentation
 
001 lpp introduction
001 lpp introduction001 lpp introduction
001 lpp introduction
 

Similar to An efficient use of temporal difference technique in Computer Game Learning

Reinforcement Learning
Reinforcement LearningReinforcement Learning
Reinforcement Learning
SVijaylakshmi
 
Reinforcement learning
Reinforcement learning Reinforcement learning
Reinforcement learning
Chandra Meena
 
reinforcement-learning-141009013546-conversion-gate02.pdf
reinforcement-learning-141009013546-conversion-gate02.pdfreinforcement-learning-141009013546-conversion-gate02.pdf
reinforcement-learning-141009013546-conversion-gate02.pdf
VaishnavGhadge1
 
Head First Reinforcement Learning
Head First Reinforcement LearningHead First Reinforcement Learning
Head First Reinforcement Learning
azzeddine chenine
 
Reinforcement learning
Reinforcement learningReinforcement learning
Reinforcement learning
DongHyun Kwak
 
Reinforcement learning
Reinforcement learningReinforcement learning
Reinforcement learning
Elias Hasnat
 
How to formulate reinforcement learning in illustrative ways
How to formulate reinforcement learning in illustrative waysHow to formulate reinforcement learning in illustrative ways
How to formulate reinforcement learning in illustrative ways
YasutoTamura1
 
RL.ppt
RL.pptRL.ppt
RL.ppt
AzharJamil15
 
REINFORCEMENT LEARNING
REINFORCEMENT LEARNINGREINFORCEMENT LEARNING
REINFORCEMENT LEARNING
pradiprahul
 
reiniforcement learning.ppt
reiniforcement learning.pptreiniforcement learning.ppt
reiniforcement learning.ppt
charusharma165
 
Introduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement LearningIntroduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement Learning
NAVER Engineering
 
Reinforcement Learning
Reinforcement LearningReinforcement Learning
Reinforcement Learning
DongHyun Kwak
 
YijueRL.ppt
YijueRL.pptYijueRL.ppt
YijueRL.ppt
Shoaib Iqbal
 
RL_online _presentation_1.ppt
RL_online _presentation_1.pptRL_online _presentation_1.ppt
RL_online _presentation_1.ppt
ssuser43a599
 
Hibridization of Reinforcement Learning Agents
Hibridization of Reinforcement Learning AgentsHibridization of Reinforcement Learning Agents
Hibridization of Reinforcement Learning Agentsbutest
 
Reinforcement Learning.ppt
Reinforcement Learning.pptReinforcement Learning.ppt
Reinforcement Learning.ppt
POOJASHREEC1
 
Reinfrocement Learning
Reinfrocement LearningReinfrocement Learning
Reinfrocement Learning
Natan Katz
 
24.09.2021 Reinforcement Learning Algorithms.pptx
24.09.2021 Reinforcement Learning Algorithms.pptx24.09.2021 Reinforcement Learning Algorithms.pptx
24.09.2021 Reinforcement Learning Algorithms.pptx
ManiMaran230751
 
Introduction to Deep Reinforcement Learning
Introduction to Deep Reinforcement LearningIntroduction to Deep Reinforcement Learning
Introduction to Deep Reinforcement Learning
IDEAS - Int'l Data Engineering and Science Association
 
Deep reinforcement learning from scratch
Deep reinforcement learning from scratchDeep reinforcement learning from scratch
Deep reinforcement learning from scratch
Jie-Han Chen
 

Similar to An efficient use of temporal difference technique in Computer Game Learning (20)

Reinforcement Learning
Reinforcement LearningReinforcement Learning
Reinforcement Learning
 
Reinforcement learning
Reinforcement learning Reinforcement learning
Reinforcement learning
 
reinforcement-learning-141009013546-conversion-gate02.pdf
reinforcement-learning-141009013546-conversion-gate02.pdfreinforcement-learning-141009013546-conversion-gate02.pdf
reinforcement-learning-141009013546-conversion-gate02.pdf
 
Head First Reinforcement Learning
Head First Reinforcement LearningHead First Reinforcement Learning
Head First Reinforcement Learning
 
Reinforcement learning
Reinforcement learningReinforcement learning
Reinforcement learning
 
Reinforcement learning
Reinforcement learningReinforcement learning
Reinforcement learning
 
How to formulate reinforcement learning in illustrative ways
How to formulate reinforcement learning in illustrative waysHow to formulate reinforcement learning in illustrative ways
How to formulate reinforcement learning in illustrative ways
 
RL.ppt
RL.pptRL.ppt
RL.ppt
 
REINFORCEMENT LEARNING
REINFORCEMENT LEARNINGREINFORCEMENT LEARNING
REINFORCEMENT LEARNING
 
reiniforcement learning.ppt
reiniforcement learning.pptreiniforcement learning.ppt
reiniforcement learning.ppt
 
Introduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement LearningIntroduction of Deep Reinforcement Learning
Introduction of Deep Reinforcement Learning
 
Reinforcement Learning
Reinforcement LearningReinforcement Learning
Reinforcement Learning
 
YijueRL.ppt
YijueRL.pptYijueRL.ppt
YijueRL.ppt
 
RL_online _presentation_1.ppt
RL_online _presentation_1.pptRL_online _presentation_1.ppt
RL_online _presentation_1.ppt
 
Hibridization of Reinforcement Learning Agents
Hibridization of Reinforcement Learning AgentsHibridization of Reinforcement Learning Agents
Hibridization of Reinforcement Learning Agents
 
Reinforcement Learning.ppt
Reinforcement Learning.pptReinforcement Learning.ppt
Reinforcement Learning.ppt
 
Reinfrocement Learning
Reinfrocement LearningReinfrocement Learning
Reinfrocement Learning
 
24.09.2021 Reinforcement Learning Algorithms.pptx
24.09.2021 Reinforcement Learning Algorithms.pptx24.09.2021 Reinforcement Learning Algorithms.pptx
24.09.2021 Reinforcement Learning Algorithms.pptx
 
Introduction to Deep Reinforcement Learning
Introduction to Deep Reinforcement LearningIntroduction to Deep Reinforcement Learning
Introduction to Deep Reinforcement Learning
 
Deep reinforcement learning from scratch
Deep reinforcement learning from scratchDeep reinforcement learning from scratch
Deep reinforcement learning from scratch
 

Recently uploaded

Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Jeffrey Haguewood
 
How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...
Product School
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
DanBrown980551
 
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdfFIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance
 
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Tobias Schneck
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
Ana-Maria Mihalceanu
 
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdfSmart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
91mobiles
 
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Ramesh Iyer
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
Alan Dix
 
The Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and SalesThe Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and Sales
Laura Byrne
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Albert Hoitingh
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
Jemma Hussein Allen
 
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMsTo Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
Paul Groth
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
James Anderson
 
Generating a custom Ruby SDK for your web service or Rails API using Smithy
Generating a custom Ruby SDK for your web service or Rails API using SmithyGenerating a custom Ruby SDK for your web service or Rails API using Smithy
Generating a custom Ruby SDK for your web service or Rails API using Smithy
g2nightmarescribd
 
Key Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdfKey Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdf
Cheryl Hung
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
OnBoard
 
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
Sri Ambati
 
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdfFIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
Guy Korland
 

Recently uploaded (20)

Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
 
How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
 
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdfFIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdf
 
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
 
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdfSmart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
 
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
 
The Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and SalesThe Art of the Pitch: WordPress Relationships and Sales
The Art of the Pitch: WordPress Relationships and Sales
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
 
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMsTo Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
 
Generating a custom Ruby SDK for your web service or Rails API using Smithy
Generating a custom Ruby SDK for your web service or Rails API using SmithyGenerating a custom Ruby SDK for your web service or Rails API using Smithy
Generating a custom Ruby SDK for your web service or Rails API using Smithy
 
Key Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdfKey Trends Shaping the Future of Infrastructure.pdf
Key Trends Shaping the Future of Infrastructure.pdf
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
 
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
GenAISummit 2024 May 28 Sri Ambati Keynote: AGI Belongs to The Community in O...
 
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdfFIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
FIDO Alliance Osaka Seminar: Passkeys at Amazon.pdf
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
 

An efficient use of temporal difference technique in Computer Game Learning

  • 1. An Efficient use of temporal difference technique in computer game learning Indian institute of technology ( Indian school of mines ), Dhanbad. Project guide:- Presented by: Dr. Rajendra Pamula Prabhu Kumar Department of computer 15MT000624 Science and engineering Computer science and engineering Indian institute of technology Indian institute of technology ( Indian school of mines), ( Indian school of mines), Dhanbad Dhanbad
  • 2. Outline 1. Introduction of reinforcement learning 2. Agent-Environment interface 3. Types of reinforcement learning 4. Elements of the reinforcement learning 5. Types of selection of state 6. Algorithms of reinforcement learning References
  • 3. Introduction of reinforcement learning  Reinforcement learning is the part of machine learning, which is a field of computer science that gives computer to ability to learn without being explicitly programmed.  Reinforcement learning is a framework for computational learning agents use experience from their interaction with an environment to improve performance over time.  In reinforcement learning task, the agent understands the state of the environment and it always tries to maximize the long-term return which is based on real value reward.  It is learning of what to do-how to do mapping situation to action so as maximize total numerical reward and minimize the penalty.
  • 4. Introduction of reinforcement learning cont.… • If there is no explicit teacher to guide the learning agent, the agent must learn the behavior through trail-and-error interaction with unknown environment. • The learning agent senses the environment, takes actions on it, and receives numeric reward or punishment from some reward function. • When we say agent learn means ”sometimes it modifies the code itself or modifies the database ”, database implies the experiences, information, event etc. • It is responsible for making decision. • The main goal of reinforcement learning is “Buildup a good model such as algorithm which generate a sequence of decision and lead to the highest long-term reward.”
  • 5. Agent environment interface o At each time step t, the reinforcement learning agent receives some representation of environment’s current state s(t) € S ,where S is the set of possible state and then choose some action a(t)€ A(st), where A(st) is set of actions that can be executed in state s(t). o The agent receives reward r(t+1) and execute in next state s(t+1) o The reward function can be used for specify the wide range of planning goals, It means the designer can tell the agent what he has achieve. o The reward function which must be unalterable by the agent.
  • 6. Types of reinforcement learning There are two types of reinforcement learning 1. Episodic: The interaction with the environment is divided into independent episodes. “Independent”, means performance in each episode is depends only the action taken on that episode. in episodic task, a return is sum of all reward received from the beginning of the episode until ends. where, T is terminal state i.e. end of episode ends S0 is the starting state of R denotes as total return r(k) denotes as the reward on the kth states
  • 7. Types of the reinforcement learning contd.. 2.Continuing task: It consist infinite sequence of state, action and rewards. In this task, the action and environment interaction doesn’t break down in separate episode. The performance is depends upon the current action. In the case of continuing task, The return is depends upon discount factor  where γ denotes discount factor which adjust the relative importance between long-term consequences vs. short term consequences.  The discount factor is between 0 and 1.  The discount factor reflects the strategy of how fast learning takes place  If γ =0, agent only concerned about maximizing the immediate rewards  If γ approaches to 1, The agent takes the future reward into account
  • 8. Element of the reinforcement learning 1. Policy: It defines the learning agent’s way of behaving at a given time. It might be a function or simple lookup table. It only used in reinforcement learning is to determine the behavior. 2. Reward function: It is the function which defines which one is the bad and good event for agent. It maps each state-action pair of the environment to a single real number. It must necessarily unalterable by the agent.
  • 9. Elements in the reinforcement learning contd.. 3. Value function: It specifies what is good in long run. The value of the state is the total amount of reward an agent can expect to accumulate over the future, starting from that state. Where as reward is the immediate desirability of environmental states. i.e. values indicate the long-tem desirability of states. 4. Model: It is used for planning. It defines the copy of behavior, e.g. by given state and action ,The model might predict the next state and next reward.
  • 10. Algorithm of reinforcement learning 1. Markov decision processes: • It is standard, general formalism for sequential decision problems. • It consist tuple of <S,A,P,R> where S is the set of states. A is the set of actions available to the agent P is the probability, P(a, ss′) = P r {st+1 = s ′ | st = s, at = a}, it is a state transition function that defines the probability of transitioning to state s ′ at time t + 1 after action a is taken when agent is in state s at time t. R is the reward function that determines the probability of receiving reward after choosing action a in state s and going for next state s’.
  • 11. Algorithms in reinforcement learning 2. Dynamic programming (DP) • It is the method to solve the markov decision process i.e. to find an optimal policy, if the full knowledge of model is available. • For dynamic programming, all the transition probabilities and reward expectation must be known. • This algorithm updates the estimates of states values based on their estimates of the next state. • There are two basic DP methods used for computing optimal policy 1. Policy iteration 2. Value iteration
  • 12. Policy Iteration: • It forms a sequence of policy Ωo, Ω1, Ω2….Ωk, Ωk+1 where Ωk+1 is an improvement of Ωk. • Policy evaluation task is concerned with computing state value function for any policy Ω • The iterative algorithm for policy evaluation is • Estimating value functions is particularly useful for finding the better policy. • The policy improvement algorithm uses action-value function to improve the current policy. If then it is better to select action a in policy Ω • If Ω and Ω’ are two policy and this condition hold then Ω’ is the better policy than Ω
  • 13. Value iteration • In value iteration, optimal policy is not computed directly. • For that, optimal value function is computed and then a greedy policy with respect to function is an optimal policy. • It stops and find the optimal policy when the changes introduced by backups/updates becomes sufficiently small. • One threshold value has been initialized and compared with threshold value. • If the policy value is sufficiently smaller than threshold, the policy is called as optimal policy
  • 14. 3. Temporal difference • The temporal differences idea has been taken from dynamic programming. • The temporal difference and dynamic programming, both are used for accumulating the value functions. • In this methods, learning takes place after every time step which is beneficial as it makes for efficient learning • The agent can revise its policy after every action and state it experiences. • TD algorithms make updates of the estimated policy values based on each state transition and on the immediate reward received from the environment on this transition. • The initial temporal difference algorithm is called TD(0),called tabular estimates v(Ω). It updates by following method V (s) ← V (s) + α (V (s’) − V (s)) where α is the positive step size parameter V(s’) is the value function for next state α (V (s’) − V (s)), called as Temporal difference error. V(s) is the value function for the current state Which is always designed to move toward 0.
  • 15. Type of selection of states 1. Greedy 2. Exploration process a) Providing initial knowledge b) Deriving a policy from demonstration c) Ask for help d) Teacher provide advice
  • 16. Application of reinforcement learning 1. Benchmark Problems a) Mountain car b) Cart-pole balancing c) Pendulum swing up tec.. 2. Games a) Tic-Tac-Toe b) Chess etc.. 3. Real world applications a) Robotics b) Control of helicopter c) Prediction of stock prices
  • 17. References • R. S. Sutton. Reinforcement learning: past, present and future [online]. Available from http: //www-anw. cs. umass. edu/ rich/Talks/SEAL98/SEAL98. html [accessed on December 2005]. 1999. • R. S. Sutton and A. G. Barto. Reinforcement learning. an introduction. Cambridge, MA: The MIT Press, 1998. • M. L. Puterman. Markov decision processes-discrete stochastic dynamic programming. John Wiley and sons, Inc, New York, NY, 1994.