Artificial Intelligence
What is an Intelligent agent?
â€ĸ “Agent” can be considered as a theoretical
concept from AI.
â€ĸ Many different definitions exist in the
literatureâ€Ļ..
Agent Definition (1)
â€ĸ An agent is an entity which is:
– Situated in some environment.
– Autonomous, in the sense that it can act without direct
intervention from humans or other software processes, and
controls over its own actions and internal state.
– Flexible which means:
â€ĸ Responsive (reactive): agents should perceive their
environment and respond to changes that occur in it;
â€ĸ Proactive: agents should not simply act in response to their
environment, they should be able to exhibit opportunistic,
goal-directed behavior and take the initiative when
appropriate;
â€ĸ Social: agents should be able to interact with humans or
other artificial agents
Agent Definition (2)
American Heritage Dictionary:
agent-
” â€Ļ one that acts or has the power or
authority to actâ€Ļ or represent
another”
â€ĸ "An agent is anything that can be viewed
as perceiving its environment through
sensors and acting upon that
environment through effectors."
Russell & Norvig
Agent Definition (3)
“Intelligent agents continuously perform three
functions:
â€ĸ perception of dynamic conditions in the
environment;
â€ĸ action to affect conditions in the environment;
and
â€ĸ reasoning to interpret perceptions, solve
problems, draw inferences, and determine
actions.”
Barbara Hayes-Roth
Agent Definition (5)
Agents & Environments
īą The agent takes sensory input from its
environment, and produces as output actions
that affect it.
Environment
sensor
input
action
outputAgent
Internal and External Environment of
an Agent
Internal Environment:
architecture, goals, abilities, sensors,
effectors, profile, knowledge,
beliefs, etc.
External Environment:
user, other humans, other agents,
applications, information sources,
their relationships,
platforms, servers, networks, etc.
Balance
Overall Intelligent Agent
[Terziyan, 1993]
1) is goal-oriented, because it should have at least one goal -
to keep continuously balance between its internal and external
environments ;
2) is creative because of the ability to change external environment;
3) is adaptive because of the ability to change internal environment;
4) is mobile because of the ability to move to another place;
5) is social because of the ability to communicate to create a
community;
6) is self-configurable because of the ability to protect “mental
health” by sensing only a “suitable” part of the environment.
Agent Definition
Wikipedia: (The free Encyclopedia),
īąIn computer science, an intelligent agent (IA) is a
software agent that exhibits some form of artificial
intelligence that assists the user and will act on their
behalf, in performing non-repetitive computer-
related tasks.
īąWhile the working of software agents used for
operator assistance or data mining is often based on
fixed pre-programmed rules, "intelligent" here
implies the ability to adapt and learn.
Three groups of agents
[Etzioni and Daniel S. Weld, 1995]
â€ĸ Backseat driver: helps the user during some task
(e.g., Microsoft Office Assistant);
â€ĸ Taxi driver: knows where to go when you tell the
destination;
â€ĸ Caretaker : know where to go, when and why.
Agent’s function maps
â€ĸ The agent function maps from percept histories to actions:
Agents example
â€ĸ Human agent:
– eyes, ears, and other organs for sensors;
– hands, legs, mouth, and other body parts for actuators
â€ĸ Robotic agent:
– cameras and infrared range finders for sensors
– various motors for actuators
PEAS
â€ĸ Use PEAS to describe task environment
– Performance measure
– Environment
– Actuators
– Sensors
â€ĸ Example: Taxi driver
– Performance measure: safe, fast, comfortable
(maximize profits)
– Environment: roads, other traffic, pedestrians,
customers
– Actuators: steering, accelerator, brake, signal, horn
– Sensors: cameras, sonar, speedometer, GPS,
odometer, accelerometer, engine sensors
PEAS
â€ĸ Agent: Part-picking robot
â€ĸ Performance measure: Percentage of parts in
correct bins
â€ĸ Environment: Conveyor belt with parts, bins
â€ĸ Actuators: Jointed arm and hand
â€ĸ Sensors: Camera, joint angle sensors
PEAS
â€ĸ Agent: Medical diagnosis system
â€ĸ Performance measure: Healthy patient, minimize
costs, lawsuits
â€ĸ Environment: Patient, hospital, staff
â€ĸ Actuators: Screen display (questions, tests,
diagnoses, treatments, referrals)
â€ĸ
â€ĸ Sensors: Keyboard (entry of symptoms, findings,
patient's answers)
PEAS
â€ĸ Agent: Interactive English tutor
â€ĸ Performance measure: Maximize student's
score on test
â€ĸ Environment: Set of students
â€ĸ Actuators: Screen display (exercises,
suggestions, corrections)
â€ĸ Sensors: Keyboard
Description of Agent‘s character
īą An agent is responsible for satisfying specific goals. There can
be different types of goals such as achieving a specific status,
maximising a given function (e.g., utility), etc.
īą The state of an agent includes state of its internal
environment + state of knowledge and beliefs about its
external environment.
knowledge
beliefs
Goal1
Goal2
Connections
â€ĸ An agent is situated in an environment, that consists of
the objects and other agents it is possible to interact
with.
â€ĸ An agent has an identity that distinguishes it from the
other agents of its environment.
environment
Agent types
â€ĸ Four basic types in order of increasing
generalization:
– Simple reflex agents
– Reflex agents with state/model
– Goal-based agents
– Utility-based agents
Simple Reflex Agent
ī‚ĸ Instead of specifying individual mappings in an explicit table,
common input-output associations are recorded
ī‚— Requires processing of percepts to achieve some
abstraction
ī‚— Frequent method of specification is through
condition-action rules
ī‚ĸ if percept then action
ī‚ĸ If car-in-front-is-braking then initiate-braking
ī‚— Similar to innate reflexes or learned responses in
humans
ī‚— Efficient implementation, but limited power
ī‚ĸ Environment must be fully observable
ī‚ĸ Easily runs into infinite loops
Simple reflex agents
Simple Reflex Agent
â€ĸ function SIMPLE-REFLEX-AGENT (percept)
returns action
– static: rules, a set of condition-action rules
– state ← INTERPRET-INPUT (percept)
– rule ← RULE-MATCH (state, rules)
– action ← RULE-ACTION [rule]
– return action
A simple reflex agent..
â€ĸ which works by finding a rule whose condition
matches the current situation and then doing
the action associated with that rule
Reflex agents with state/model
â€ĸ Evan a little bit of un observability can cause
serious trouble.
– The braking rule given earlier assumes that the
condition car-in-front-is-braking can be
determined from the current percept – the
current video image.
â€ĸ More advanced techniques would require the
maintenance of some kind of internal state to
choose an action.
Reflex agents with state/model
ī‚ĸ An internal state maintains important information from
previous percepts
ī‚— Sensors only provide a partial picture of the environment
ī‚— Helps with some partially observable environments
ī‚ĸ The internal states reflects the agent’s knowledge about
the world
ī‚— This knowledge is called a model
ī‚— May contain information about changes in the world
Model-based reflex agents
â€ĸ Required information:
– How the world evolves independently of the
agent?
â€ĸ An overtaking car generally will be closer behind than it
was a moment ago.
â€ĸ The current percept is combined with the old internal
state to generate the updated description of the
current state.
Model-based reflex agents
Model-based reflex agents
â€ĸ function REFLEX-AGENT-WITH-STATE (percept)
returns an action
– static: state, a description of the current world state
rules, a set of condition-action rules
action, the most recent action, initially none
– state ← UPDATE-STATE (state, action, percept)
– rule ← RULE-MATCH (state, rules)
– action ← RULE-ACTION [rule]
– state ← UPDATE-STATE (state, action)
– return action
Goal-based agent
â€ĸ Merely knowing about the current state of the
environment is not always enough to decide
what to do next.
â€ĸ The right decision depends on where the taxi
is trying to get to.
â€ĸ So the goal information is also needed.
Goal-based agent
â€ĸ Goal-based agents are far more flexible.
– If it starts to rain, the agent adjusts itself to the
changed circumstances, since it also looks at the
way its actions would affect its goals (remember
doing the right thing).
– For the reflex agent we would have to rewrite a
large number of condition-action rules.
Goal-based agents
Utility-based agents
â€ĸ Goals are not really enough to generate high-
quality behavior.
â€ĸ There are many ways to reach the destination,
but some are qualitatively better than others.
– More safe
– Shorter
– Less expensive
Utility-based agent
â€ĸ We say that if one world state is preferred to
another, then it has higher utility for the
agent.
â€ĸ Utility is a function that maps a state onto a
real number.
– state → R
â€ĸ Any rational agent possesses a utility function.
Utility-based agents
Environments
â€ĸ Actions are done by the agent on the environment.
â€ĸ Environment provides percepts to the agent.
â€ĸ Determine to a large degree the interaction between
the “outside world” and the agent
– the “outside world” is not necessarily the “real world” as
we perceive it
â€ĸ it may be a real or virtual environment the agent lives in
â€ĸ In many cases, environments are implemented within
computers
– They may or may not have a close correspondence to the
“real world”
Properties of environments
â€ĸ Fully observable vs. partially observable
â€ĸ Or Accessible vs. inaccessible
– If an agent’s sensory equipment gives it access to the
complete state of the environment, then we say that
environment is fully observable to the agent.
– An environment is effectively fully observable if the
sensors detect all aspects that are relevant to the choice of
action.
– A fully observable environment is convenient because the
agent need not maintain any internal state to keep track of
the world.
Properties of environments
â€ĸ Deterministic vs. nondeterministic.
– If the next state of the environment is completely
determined by the current state and the actions
selected by the agents, then we say the
environment is deterministic.
– If the environment is inaccessible, then it may
appear to be nondeterministic (bunch of
uncertainties).
Properties of task environments
â€ĸ Episodic vs. sequential.
– Agent’s experience is divided into “episodes.”
â€ĸ Each episode consists of the agent perceiving and
acting.
– Subsequent episodes do not depend on what
actions occur in previous episodes.
– In sequential environments current actions affect
all succeeding actions
Properties of task environments
â€ĸ Static vs. Dynamic
– If the environment can change while an agent is
performing action, then we say the environment is
dynamic.
– Else its static.
– Static environments are easy to deal with, because the
agent does not keep on looking at the environment
while it is deciding on an action.
– Semidynamic: if the environment does not change
with the passage of time but the agent performance
score does.
Properties of environments
â€ĸ Discrete vs. continuous
– If there are a limited number of distinct, clearly
defined percepts and actions, we say that the
environment is discrete.
â€ĸ Chess, since there are a fixed number of possible
moves on each turn.
â€ĸ Taxi driving is continuous.
Properties of environments
â€ĸ Single agent vs. Multiagent
– In the single agent environment there is only one
agent
â€ĸ A computer software playing crossword puzzle
– In multiagent systems, there are more than one
active agents
â€ĸ Video games
Environment Obser
vable
Determi
nistic
Episodic Stati
c
Discrete Agents
Chess with a clock
Chess without a clock
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
EnvironmentExamples
EnvironmentExamples
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Stati
c
Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Stati
c
Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Stati
c
Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis Partial Stochast
ic
Episodic Static Continu
ous
Single
Environment Examples
Fully observable vs. partially observable
Deterministic vs. stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs. multiagent
Environment Obser
vable
Determi
nistic
Episodic Stati
c
Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis Partial Stochast
ic
Episodic Static Continu
ous
Single
Image analysis
Environment Examples
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis Partial Stochast
ic
Episodic Static Continu
ous
Single
Image analysis Fully Determi
nistic
Episodic Semi Discrete Single
Fully observable vs.
partially observable
Deterministic vs.
stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs.
multiagent
Environment Examples
Fully observable vs.
partially observable
Deterministic vs.
stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs.
multiagent
Environment Obser
vable
Determi
nistic
Episodic Stati
c
Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis Partial Stochast
ic
Episodic Static Continu
ous
Single
Image analysis Fully Determi
nistic
Episodic Semi Discrete Single
Robot part picking
Environment Examples
Fully observable vs.
partially observable
Deterministic vs.
stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs.
multiagent
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis Partial Stochast
ic
Episodic Static Continu
ous
Single
Image analysis Fully Determi
nistic
Episodic Semi Discrete Single
Robot part picking Fully Determi
nistic
Episodic Semi Discrete Single
Environment Examples
Fully observable vs.
partially observable
Deterministic vs.
stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs.
multiagent
Environment Obser
vable
Determi
nistic
Episodic Stati
c
Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis Partial Stochast
ic
Episodic Static Continu
ous
Single
Image analysis Fully Determi
nistic
Episodic Semi Discrete Single
Robot part picking Fully Determi
nistic
Episodic Semi Discrete Single
Interactive English
tutor
Environment Examples
Fully observable vs.
partially observable
Deterministic vs.
stochastic / strategic
Episodic vs. sequential
Static vs. dynamic
Discrete vs. continuous
Single agent vs.
multiagent
Environment Obser
vable
Determi
nistic
Episodic Static Discrete Agents
Chess with a clock Fully Strategic Sequential Semi Discrete Multi
Chess without a clock Fully Strategic Sequential Static Discrete Multi
Poker Partial Strategic Sequential Static Discrete Multi
Backgammon Fully Stochast
ic
Sequential Static Discrete Multi
Taxi driving Partial Stochast
ic
Sequential Dyna
mic
Continu
ous
Multi
Medical diagnosis Partial Stochast
ic
Episodic Static Continu
ous
Single
Image analysis Fully Determi
nistic
Episodic Semi Discrete Single
Robot part picking Fully Determi
nistic
Episodic Semi Discrete Single
Interactive English
tutor
Partial Stochast
ic
Sequential Dyna
mic
Discrete Multi

Artificial intelligence(03)

  • 1.
  • 2.
    What is anIntelligent agent? â€ĸ “Agent” can be considered as a theoretical concept from AI. â€ĸ Many different definitions exist in the literatureâ€Ļ..
  • 3.
    Agent Definition (1) â€ĸAn agent is an entity which is: – Situated in some environment. – Autonomous, in the sense that it can act without direct intervention from humans or other software processes, and controls over its own actions and internal state. – Flexible which means: â€ĸ Responsive (reactive): agents should perceive their environment and respond to changes that occur in it; â€ĸ Proactive: agents should not simply act in response to their environment, they should be able to exhibit opportunistic, goal-directed behavior and take the initiative when appropriate; â€ĸ Social: agents should be able to interact with humans or other artificial agents
  • 4.
    Agent Definition (2) AmericanHeritage Dictionary: agent- ” â€Ļ one that acts or has the power or authority to actâ€Ļ or represent another”
  • 5.
    â€ĸ "An agentis anything that can be viewed as perceiving its environment through sensors and acting upon that environment through effectors." Russell & Norvig Agent Definition (3)
  • 6.
    “Intelligent agents continuouslyperform three functions: â€ĸ perception of dynamic conditions in the environment; â€ĸ action to affect conditions in the environment; and â€ĸ reasoning to interpret perceptions, solve problems, draw inferences, and determine actions.” Barbara Hayes-Roth Agent Definition (5)
  • 7.
    Agents & Environments īąThe agent takes sensory input from its environment, and produces as output actions that affect it. Environment sensor input action outputAgent
  • 8.
    Internal and ExternalEnvironment of an Agent Internal Environment: architecture, goals, abilities, sensors, effectors, profile, knowledge, beliefs, etc. External Environment: user, other humans, other agents, applications, information sources, their relationships, platforms, servers, networks, etc. Balance
  • 9.
    Overall Intelligent Agent [Terziyan,1993] 1) is goal-oriented, because it should have at least one goal - to keep continuously balance between its internal and external environments ; 2) is creative because of the ability to change external environment; 3) is adaptive because of the ability to change internal environment; 4) is mobile because of the ability to move to another place; 5) is social because of the ability to communicate to create a community; 6) is self-configurable because of the ability to protect “mental health” by sensing only a “suitable” part of the environment.
  • 10.
    Agent Definition Wikipedia: (Thefree Encyclopedia), īąIn computer science, an intelligent agent (IA) is a software agent that exhibits some form of artificial intelligence that assists the user and will act on their behalf, in performing non-repetitive computer- related tasks. īąWhile the working of software agents used for operator assistance or data mining is often based on fixed pre-programmed rules, "intelligent" here implies the ability to adapt and learn.
  • 11.
    Three groups ofagents [Etzioni and Daniel S. Weld, 1995] â€ĸ Backseat driver: helps the user during some task (e.g., Microsoft Office Assistant); â€ĸ Taxi driver: knows where to go when you tell the destination; â€ĸ Caretaker : know where to go, when and why.
  • 12.
    Agent’s function maps â€ĸThe agent function maps from percept histories to actions:
  • 13.
    Agents example â€ĸ Humanagent: – eyes, ears, and other organs for sensors; – hands, legs, mouth, and other body parts for actuators â€ĸ Robotic agent: – cameras and infrared range finders for sensors – various motors for actuators
  • 14.
    PEAS â€ĸ Use PEASto describe task environment – Performance measure – Environment – Actuators – Sensors â€ĸ Example: Taxi driver – Performance measure: safe, fast, comfortable (maximize profits) – Environment: roads, other traffic, pedestrians, customers – Actuators: steering, accelerator, brake, signal, horn – Sensors: cameras, sonar, speedometer, GPS, odometer, accelerometer, engine sensors
  • 15.
    PEAS â€ĸ Agent: Part-pickingrobot â€ĸ Performance measure: Percentage of parts in correct bins â€ĸ Environment: Conveyor belt with parts, bins â€ĸ Actuators: Jointed arm and hand â€ĸ Sensors: Camera, joint angle sensors
  • 16.
    PEAS â€ĸ Agent: Medicaldiagnosis system â€ĸ Performance measure: Healthy patient, minimize costs, lawsuits â€ĸ Environment: Patient, hospital, staff â€ĸ Actuators: Screen display (questions, tests, diagnoses, treatments, referrals) â€ĸ â€ĸ Sensors: Keyboard (entry of symptoms, findings, patient's answers)
  • 17.
    PEAS â€ĸ Agent: InteractiveEnglish tutor â€ĸ Performance measure: Maximize student's score on test â€ĸ Environment: Set of students â€ĸ Actuators: Screen display (exercises, suggestions, corrections) â€ĸ Sensors: Keyboard
  • 18.
    Description of Agent‘scharacter īą An agent is responsible for satisfying specific goals. There can be different types of goals such as achieving a specific status, maximising a given function (e.g., utility), etc. īą The state of an agent includes state of its internal environment + state of knowledge and beliefs about its external environment. knowledge beliefs Goal1 Goal2
  • 19.
    Connections â€ĸ An agentis situated in an environment, that consists of the objects and other agents it is possible to interact with. â€ĸ An agent has an identity that distinguishes it from the other agents of its environment. environment
  • 20.
    Agent types â€ĸ Fourbasic types in order of increasing generalization: – Simple reflex agents – Reflex agents with state/model – Goal-based agents – Utility-based agents
  • 21.
    Simple Reflex Agent ī‚ĸInstead of specifying individual mappings in an explicit table, common input-output associations are recorded ī‚— Requires processing of percepts to achieve some abstraction ī‚— Frequent method of specification is through condition-action rules ī‚ĸ if percept then action ī‚ĸ If car-in-front-is-braking then initiate-braking ī‚— Similar to innate reflexes or learned responses in humans ī‚— Efficient implementation, but limited power ī‚ĸ Environment must be fully observable ī‚ĸ Easily runs into infinite loops
  • 22.
  • 23.
    Simple Reflex Agent â€ĸfunction SIMPLE-REFLEX-AGENT (percept) returns action – static: rules, a set of condition-action rules – state ← INTERPRET-INPUT (percept) – rule ← RULE-MATCH (state, rules) – action ← RULE-ACTION [rule] – return action
  • 24.
    A simple reflexagent.. â€ĸ which works by finding a rule whose condition matches the current situation and then doing the action associated with that rule
  • 25.
    Reflex agents withstate/model â€ĸ Evan a little bit of un observability can cause serious trouble. – The braking rule given earlier assumes that the condition car-in-front-is-braking can be determined from the current percept – the current video image. â€ĸ More advanced techniques would require the maintenance of some kind of internal state to choose an action.
  • 26.
    Reflex agents withstate/model ī‚ĸ An internal state maintains important information from previous percepts ī‚— Sensors only provide a partial picture of the environment ī‚— Helps with some partially observable environments ī‚ĸ The internal states reflects the agent’s knowledge about the world ī‚— This knowledge is called a model ī‚— May contain information about changes in the world
  • 27.
    Model-based reflex agents â€ĸRequired information: – How the world evolves independently of the agent? â€ĸ An overtaking car generally will be closer behind than it was a moment ago. â€ĸ The current percept is combined with the old internal state to generate the updated description of the current state.
  • 28.
  • 29.
    Model-based reflex agents â€ĸfunction REFLEX-AGENT-WITH-STATE (percept) returns an action – static: state, a description of the current world state rules, a set of condition-action rules action, the most recent action, initially none – state ← UPDATE-STATE (state, action, percept) – rule ← RULE-MATCH (state, rules) – action ← RULE-ACTION [rule] – state ← UPDATE-STATE (state, action) – return action
  • 30.
    Goal-based agent â€ĸ Merelyknowing about the current state of the environment is not always enough to decide what to do next. â€ĸ The right decision depends on where the taxi is trying to get to. â€ĸ So the goal information is also needed.
  • 31.
    Goal-based agent â€ĸ Goal-basedagents are far more flexible. – If it starts to rain, the agent adjusts itself to the changed circumstances, since it also looks at the way its actions would affect its goals (remember doing the right thing). – For the reflex agent we would have to rewrite a large number of condition-action rules.
  • 32.
  • 33.
    Utility-based agents â€ĸ Goalsare not really enough to generate high- quality behavior. â€ĸ There are many ways to reach the destination, but some are qualitatively better than others. – More safe – Shorter – Less expensive
  • 34.
    Utility-based agent â€ĸ Wesay that if one world state is preferred to another, then it has higher utility for the agent. â€ĸ Utility is a function that maps a state onto a real number. – state → R â€ĸ Any rational agent possesses a utility function.
  • 35.
  • 36.
    Environments â€ĸ Actions aredone by the agent on the environment. â€ĸ Environment provides percepts to the agent. â€ĸ Determine to a large degree the interaction between the “outside world” and the agent – the “outside world” is not necessarily the “real world” as we perceive it â€ĸ it may be a real or virtual environment the agent lives in â€ĸ In many cases, environments are implemented within computers – They may or may not have a close correspondence to the “real world”
  • 37.
    Properties of environments â€ĸFully observable vs. partially observable â€ĸ Or Accessible vs. inaccessible – If an agent’s sensory equipment gives it access to the complete state of the environment, then we say that environment is fully observable to the agent. – An environment is effectively fully observable if the sensors detect all aspects that are relevant to the choice of action. – A fully observable environment is convenient because the agent need not maintain any internal state to keep track of the world.
  • 38.
    Properties of environments â€ĸDeterministic vs. nondeterministic. – If the next state of the environment is completely determined by the current state and the actions selected by the agents, then we say the environment is deterministic. – If the environment is inaccessible, then it may appear to be nondeterministic (bunch of uncertainties).
  • 39.
    Properties of taskenvironments â€ĸ Episodic vs. sequential. – Agent’s experience is divided into “episodes.” â€ĸ Each episode consists of the agent perceiving and acting. – Subsequent episodes do not depend on what actions occur in previous episodes. – In sequential environments current actions affect all succeeding actions
  • 40.
    Properties of taskenvironments â€ĸ Static vs. Dynamic – If the environment can change while an agent is performing action, then we say the environment is dynamic. – Else its static. – Static environments are easy to deal with, because the agent does not keep on looking at the environment while it is deciding on an action. – Semidynamic: if the environment does not change with the passage of time but the agent performance score does.
  • 41.
    Properties of environments â€ĸDiscrete vs. continuous – If there are a limited number of distinct, clearly defined percepts and actions, we say that the environment is discrete. â€ĸ Chess, since there are a fixed number of possible moves on each turn. â€ĸ Taxi driving is continuous.
  • 42.
    Properties of environments â€ĸSingle agent vs. Multiagent – In the single agent environment there is only one agent â€ĸ A computer software playing crossword puzzle – In multiagent systems, there are more than one active agents â€ĸ Video games
  • 43.
    Environment Obser vable Determi nistic Episodic Stati c DiscreteAgents Chess with a clock Chess without a clock Fully observable vs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent EnvironmentExamples
  • 44.
    EnvironmentExamples Environment Obser vable Determi nistic Episodic StaticDiscrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Fully observable vs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent
  • 45.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Stati c Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker
  • 46.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Static Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi
  • 47.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Stati c Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon
  • 48.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Static Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi
  • 49.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Static Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi
  • 50.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Stati c Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis
  • 51.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Static Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis Partial Stochast ic Episodic Static Continu ous Single
  • 52.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Stati c Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis Partial Stochast ic Episodic Static Continu ous Single Image analysis
  • 53.
    Environment Examples Environment Obser vable Determi nistic EpisodicStatic Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis Partial Stochast ic Episodic Static Continu ous Single Image analysis Fully Determi nistic Episodic Semi Discrete Single Fully observable vs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent
  • 54.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Stati c Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis Partial Stochast ic Episodic Static Continu ous Single Image analysis Fully Determi nistic Episodic Semi Discrete Single Robot part picking
  • 55.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Static Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis Partial Stochast ic Episodic Static Continu ous Single Image analysis Fully Determi nistic Episodic Semi Discrete Single Robot part picking Fully Determi nistic Episodic Semi Discrete Single
  • 56.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Stati c Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis Partial Stochast ic Episodic Static Continu ous Single Image analysis Fully Determi nistic Episodic Semi Discrete Single Robot part picking Fully Determi nistic Episodic Semi Discrete Single Interactive English tutor
  • 57.
    Environment Examples Fully observablevs. partially observable Deterministic vs. stochastic / strategic Episodic vs. sequential Static vs. dynamic Discrete vs. continuous Single agent vs. multiagent Environment Obser vable Determi nistic Episodic Static Discrete Agents Chess with a clock Fully Strategic Sequential Semi Discrete Multi Chess without a clock Fully Strategic Sequential Static Discrete Multi Poker Partial Strategic Sequential Static Discrete Multi Backgammon Fully Stochast ic Sequential Static Discrete Multi Taxi driving Partial Stochast ic Sequential Dyna mic Continu ous Multi Medical diagnosis Partial Stochast ic Episodic Static Continu ous Single Image analysis Fully Determi nistic Episodic Semi Discrete Single Robot part picking Fully Determi nistic Episodic Semi Discrete Single Interactive English tutor Partial Stochast ic Sequential Dyna mic Discrete Multi