Professional Documents
Culture Documents
Intelligent Agents
Outline
Agents and environments
Rationality
PEAS (Performance measure,
Environment, Actuators, Sensors)
Environment types
Agent types
Agents
An agent is anything that can be viewed as
perceiving its environment through sensors and
acting upon that environment through actuators.
Human agent has, for example:
eyes, ears, and other organs as sensors;
hands, legs, mouth, and other body parts as
actuators.
Robotic agent has, for example:
cameras and infrared range finders as sensors;
various motors as actuators.
Agents
A software agent receives keystrokes, file
contents, and network packets as sensory
inputs and acts on the environment by
displaying on the screen, writing files, and
sending network packets.
Agent and Environment
The Vacuum-Cleaner Mini-World
Percept Action
sequences
[A,Clean] Right
[A, Dirty] Suck
[B, Clean] Left
[B, Dirty] Suck
[A, Dirty], [A, Clean] Right
[A, Clean], [B, Dirty] Suck
[A, Clean], [B, Clean] No-op
... ...
Agent Function
The agent function maps from percept histories to
actions:
[f: P* A]
An agent is completely specified by the agent
function mapping percept sequences to actions
The agent program runs on the physical
architecture to produce f.
agent = architecture + program
The Vacuum-Cleaner Mini-World
costs, lawsuits
Environment: Patient, hospital, staff
treatments, referrals.
Sensors: Keyboard entry of symptoms, findings,
patient's answers
PEAS
Agent: Part-picking robot
Performance measure: Percentage of parts in
correct bins
Environment: Conveyor belt with parts, bins
on test
Environment: Set of students, Testing Agency
Episodic?? NO NO NO NO
Single-agent?? YES NO NO NO
Examples of Environment Types
Output = action
function REFLEX-AGENT-WITH-STATE(percept)
returns an action
static: rules, a set of condition-action rules
state, a description of the current world
state
(action, the most recent action)
state UPDATE-STATE(state, (action,) percept)
rule RULE-MATCH(state, rule)
action RULE-ACTION[rule]
return action
Goal-based Agents
The agent needs a goal to
know which situations are
desirable.
Things become difficult when long
sequences of actions are required
to reach the goal.
Typically investigated in search
and planning research.
Major difference: future is taken
into account.
Is more flexible since
knowledge is represented
explicitly - to a certain degree -
and can be manipulated.
Utility-based Agents
Certain goals can be reached
in different ways.
Some are better, have a
higher utility.
Utility function maps a
(sequence of) state(s) onto a
real number.
Improvement on goal setting:
Selecting between conflicting
goals.
Select appropriately between
several goals based on
likelihood of success.
Utility-based Agents
Utility-based agent act based not only goals
but also the best way to achieve the goal.
The Utility-based agent is useful when
there are multiple possible alternatives, and
an agent has to choose in order to perform
the best action.
The utility function maps each state to a
real number to check how efficiently each
action achieves the goals.
Learning Agents
A learning agent in AI is the
type of agent which can
learn from its past
experiences, or it has
learning capabilities.