5. For passing this test, requirements are:
Natural Language Processing
Knowledge Representation
Automated Reasoning
Machine Learning
If physical objects are involved:
Computer Vision
Robotics
6. Cognitive Modeling Approach
Requires scientific theories of internal activities of
the brain
Validation is done by:
Predicting and testing behavior of human subjects
(top-down approach) – cognitive science
Direct identification from Neurological data
(bottom-up approach) – cognitive neuroscience
7. Laws of Thought Approach
Aristotle: What are correct arguments/thought
processes?
Several Greek schools developed various forms of
logic: notation and rules of derivation for thoughts;
may or may not have proceeded to the idea of
mechanization
E.g.
Socrates is a man.
All men are mortal.
=> Socrates is mortal.
8. The Rational Agent Approach
Rational behavior: doing the right thing
The right thing: that which is expected to
maximize goal achievement, given the
available information
Doesn't necessarily involve thinking – e.g.,
Recoiling from a hot stove is a reflex action that
is usually more successful than a slower action
taken after careful deliberation
9. Philosophy Logic, methods of reasoning, mind as physical
system foundations of learning, language,
rationality
Mathematics Formal representation and proof algorithms,
computation, (un)decidability, (in)tractability,
probability
Economics utility, decision theory
Neuroscience physical substrate for mental activity
Psychology phenomena of perception and experimental
techniques
Computer building fast computers
engineering
Control theory design systems that maximize an objective
function over time
Linguistics knowledge representation, natural language
processing
10. Autonomous Planning & Scheduling
Game Playing
Autonomous Control
Diagnosis
Logistics Planning
Robotics
Language Understanding & Problem Solving
11. An agent is anything that can be viewed as
perceiving its environment through sensors
and acting upon that environment through
actuators.
Human agent:
Sensors – Eyes, ears, nose, etc.
Actuators – Hands, legs, mouth, etc.
Robotic agent:
Sensors – Cameras, Infrared range finders
Actuators – Various Motors
12. An agent’s perceived choice of action at any
given instant can depend on the current
percept or the entire percept sequence .
Every agent’s external behavior is described by
the agent function and internal characterization
is done by agent program.
13. The agent function maps from percept histories to
actions:
[f: P* A]
The agent program runs on the physical
architecture to produce f
Agent = Architecture +Program
14. An agent should strive to "do the right thing",
based on what it can perceive and the actions it
can perform. The right action is the one that
will cause the agent to be most successful
Performance measure: An objective criterion
for success of an agent's behavior
E.g., performance measure of a vacuum-cleaner
agent could be amount of dirt cleaned up,
amount of time taken, amount of electricity
consumed, amount of noise generated, etc.
15. Rational Agent: For each possible percept
sequence, a rational agent should select an
action that is expected to maximize its
performance measure, given the evidence
provided by the percept sequence and
whatever built-in knowledge the agent has.
16. Rationality of an agent is determined by:
Performance Measure
Prior Knowledge of Environment
Actions ( Actuators )
Percept Sequence ( Sensors )
Artificial Intelligent Agent = initial knowledge +
ability to learn
17. Consider the task of designing an automated
taxi driver:
Performance measure: Safe, fast, legal, comfortable
trip, maximize profits
Environment: Roads, other traffic, pedestrians,
customers
Actuators: Steering wheel, accelerator, brake, horn
Sensors: Cameras, sonar, speedometer, GPS, engine
sensors, etc.
19. Agent: Part-picking robot
Performance measure: Percentage of parts in correct
bins
Environment: Conveyor belt with parts, bins
Actuators: Jointed arm and hand
Sensors: Camera, joint angle sensors
20. Fully observable (vs. partially observable): An agent's
sensors give it access to the complete state of the
environment at each point in time.
Deterministic (vs. stochastic): The next state of the
environment is completely determined by the current
state and the action executed by the agent. (If the
environment is deterministic except for the actions of
other agents, then the environment is strategic)
Episodic (vs. sequential): The agent's experience is
divided into atomic "episodes" (each episode consists
of the agent perceiving and then performing a single
action), and the choice of action in each episode
depends only on the episode itself.
21. Static (vs. dynamic): The environment is
unchanged while an agent is deliberating. (The
environment is semi dynamic if the
environment itself does not change with the
passage of time but the agent's performance
score does)
Discrete (vs. continuous): A limited number of
distinct, clearly defined percepts and actions.
Single agent (vs. multi agent): An agent
operating by itself in an environment.
22. Agent = Architecture +Program
Job of AI = design agent program which
implements the agent function mapping
percepts to actions
23. Four basic types in order of increasing
generality:
Simple reflex agents
Model-based reflex agents
Goal-based agents
Utility-based agents