intelligent agents - sharif

43
Intelligent Agents CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2016 “Artificial Intelligence: A Modern Approach”, Chapter 2 Soleymani

Upload: others

Post on 15-Oct-2021

8 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Intelligent Agents - Sharif

Intelligent AgentsCE417: Introduction to Artificial Intelligence

Sharif University of Technology

Spring 2016

“Artificial Intelligence: A Modern Approach”, Chapter 2

Soleymani

Page 2: Intelligent Agents - Sharif

Outline

Agents and environments

Rationality

PEAS (Performance measure, Environment, Actuators,

Sensors)

Environment types

Agent types

2

Page 3: Intelligent Agents - Sharif

Agents

An agent is anything that can be viewed as Sensors: perceive environment

Actuators: act upon environment

Samples of agents Human agent

Sensors: eyes, ears, and other organs for sensors

Actuators: hands, legs, vocal tract, and other movable or changeable bodyparts

Robotic agent

Sensors: cameras and infrared range finders

Actuators: various motors

Software agents

Sensors: keystrokes, file contents, received network packages

Actuators: displays on the screen, files, sent network packets

3

Page 4: Intelligent Agents - Sharif

Agents & environments

Agent behavior can be described as an agent function that

maps entire perception histories to actions:

𝑓: 𝑃∗ 𝐴

The agent program runs on the physical architecture to

produce f

Program is a concrete implementation of agent function

Architecture includes sensors, actuators, computing device

agent = architecture + program

Percept sequence to dateAction set

4

Page 5: Intelligent Agents - Sharif

Vacuum-cleaner world

Percepts: location and dirt/clean status of its location

e.g., [A,Dirty]

Actions: Left, Right, Suck,NoOp

5

Page 6: Intelligent Agents - Sharif

A vacuum-cleaner agent

One simple rule implementing the agent function:

If the current square is dirty then suck, otherwise move to the other square

Tabulation of the agent function

Percept Sequence Action

[A, Clean] Right

[A, Dirty] Suck

[B, Clean] Left

[ B, Dirty] Suck

[A, Clean], [A, Clean] Right

[A, Clean], [A, Dirty] Suck

… …

[A, Clean], [A, Clean], [A, Clean] Right

[A, Clean], [A, Clean], [A, Dirty] Suck

… …

6

Page 7: Intelligent Agents - Sharif

Rational agents

"do the right thing" based on the perception history and theactions it can perform.

Rational Agent: For each possible percept sequence, arational agent should select an action that is expected tomaximize its performance measure, given the evidenceprovided by the percept sequence and whatever built-inknowledge the agent has.

7

Page 8: Intelligent Agents - Sharif

Performance measure

Evaluates the sequence of environment states

Vacuum-cleaner agent: samples of performance measure

Amount of dirt cleaned up

One point award for each clean square at each time step

Penalty for electricity consumption & generated noise

Mediocre job or periods of high and low activation?

8

Page 9: Intelligent Agents - Sharif

Rational agents (vacuum cleaner example)

Is this rational? If dirty then suck, otherwise move to the

other square

Depends on

Performance measure, e.g., Penalty for energy consumption?

Environment, e.g., New dirt can appear?

Actuators, e.g., No-op action?

Sensors, e.g., Only sense dirt in its location?

9

Page 10: Intelligent Agents - Sharif

Rationality vs. Omniscience

Rationality is distinct from omniscience (all-knowing with

infinite knowledge, impossible in reality)

Doing actions in order to modify future percepts to

obtain useful information

information gathering or exploration (important for rationality)

e.g., eyeballs and/or neck movement in human to see different

directions

10

Page 11: Intelligent Agents - Sharif

Autonomy

An agent is autonomous if its behavior is determined by its

own experience (with ability to learn and adapt)

Not just relies only on prior knowledge of designer

Learns to compensate for partial or incorrect prior knowledge

Benefit: changing environment

Starts by acting randomly or base on designer knowledge and then learns

form experience

Rational agent should be autonomous

Example: vacuum-cleaner agent

If dirty then suck, otherwise move to the other square

Does it yield an autonomous agent?

learning to foresee occurrence of dirt in squares

11

Page 12: Intelligent Agents - Sharif

Task Environment (PEAS)

Performance measure

Environment

Actuators

Sensors

12

Page 13: Intelligent Agents - Sharif

PEAS Samples…

Agent:Automated taxi driver

Performance measure: Safe, fast, legal, comfortable trip,

maximize profits,…

Environment: Roads, other traffic, pedestrians, customers,…

Actuators: Steering wheel, accelerator, brake, signal, horn,

display

Sensors: Cameras, sonar, speedometer, GPS, odometer,

accelerometer, engine sensors, keyboard

13

Page 14: Intelligent Agents - Sharif

PEAS Samples…

Agent: Medical diagnosis system

Performance measure: Healthy patient, minimize costs

Environment: Patient, hospital, staff

Actuators: Screen display (questions, tests, diagnoses,

treatments, referrals)

Sensors: Keyboard (entry of symptoms, findings, patient's

answers)

14

Page 15: Intelligent Agents - Sharif

PEAS Samples…

Satellite image analysis system

Performance measure: Correct image categorization

Environment: Downlink from orbiting satellite

Actuators: Display of scene categorization

Sensors: Color pixel array

15

Page 16: Intelligent Agents - Sharif

PEAS Samples…

Agent: Part picking robot

Performance measure: Percentage of parts in correct bins

Environment: Conveyor belt with parts, bins

Actuators: Jointed arm and hand

Sensors: Camera, joint angle sensors

16

Page 17: Intelligent Agents - Sharif

PEAS Samples…

Agent: Interactive English tutor

Performance measure: Maximize student's score on test

Environment: Set of students

Actuators: Screen display (exercises, suggestions,corrections)

Sensors: Keyboard

17

Page 18: Intelligent Agents - Sharif

PEAS Samples…

Agent: Pacman Performance measure: Score, lives

Environment: Maze containing white dots, four ghosts, powerpills, occasionally appearing fruit

Actuators:Arrow keys

Sensors: Game screen

18

Page 19: Intelligent Agents - Sharif

Environment types

Fully observable (vs. partially observable): Sensors give access

to the complete state of the environment at each time

Sensors detect all aspects relevant to the choice of action

Convenient (need not any internal state)

Noisy and inaccurate sensors or missing parts of the state from

sensors cause partially observability

19

Page 20: Intelligent Agents - Sharif

Environment types

Deterministic (vs. stochastic): Next state can be

completely determined by the current state and the

executed action If the environment is deterministic except for the actions of

other agents, then the environment is strategic (we ignore this

uncertainty)

Partially observable environment could appear to be

stochastic.

Environment is uncertain if it is not fully observable or not

deterministic

20

Page 21: Intelligent Agents - Sharif

Environment types

21

Single agent (vs. multi-agent):

Crossword puzzle is a single-agent game (chess is a multi-agent one)

Is B an agent or just an object in the environment?

B is an agent when its behavior can be described as maximizing a performance

measure whose value depends on A’s behavior.

Multi-agent: competitive, cooperative

Randomized behavior and communication can be rational

Discrete (vs. continuous): A limited number of distinct, clearly

defined states, percepts and actions, time steps

Chess has finite number of discrete states, and discrete set of percepts

and actions while Taxi driving has continuous states, and actions

Page 22: Intelligent Agents - Sharif

Environment types

Episodic (vs. sequential): The agent's experience is divided intoatomic "episodes“ where the choice of action in each episodedepends only on the episode itself. E.g., spotting defective parts on an assembly line (independency)

In sequential environments, short-term actions can have long-term

consequences

Episodic environment can be much simpler

Static (vs. dynamic): The environment is unchanged while anagent is deliberating.

Semi-dynamic: if the environment itself does not change with thepassage of time but the agent's performance score does.

Static (cross-word puzzles), dynamic (taxi driver), semi-dynamic(clock chess)

22

Page 23: Intelligent Agents - Sharif

Environment types

Known (vs. unknown): the outcomes or (outcomes

probabilities for all actions are given. It is not strictly a property of the environment

Related to agent’s or designer’s state of knowledge about “laws of physics” of

the environment

The real world is partially observable, multi-agent, stochastic,

sequential, dynamic, continuous, (and unknown)

Hardest type of environment

The environment type largely determines the agent design

23

Page 24: Intelligent Agents - Sharif

Pacman game

Fully observable?

Single-agent?

Deterministic?

Discrete?

Episodic?

Static?

Known?

24

Page 25: Intelligent Agents - Sharif

Environment types

25

Page 26: Intelligent Agents - Sharif

Environment types

26

Page 27: Intelligent Agents - Sharif

Structure of agents

An agent is completely specified by the agent function (that

maps percept sequences to actions)

One agent function or small equivalent class is rational

Agent program implements agent function (focus of our

course)

Agent program takes just the current percept as input

Agent needs to remember the whole percept sequence, if requiring it

(internal state)

27

Page 28: Intelligent Agents - Sharif

Agent Program Types

Lookup table

Basic types of agent program in order of increasinggenerality:

Simple reflexive

Model-based reflex agents

Goal-based agents

Utility-based agents

Learning-based agents

28

We mainly focus on this type in our course

Page 29: Intelligent Agents - Sharif

Look Up Table Agents

Benefits:

Easy to implement

Drawbacks:

space ( 𝑡=1𝑇 |𝑃|t;𝑃 : set of possible percepts, 𝑇: lifetime)

For chess at least 10150 entries while less than 1080 atoms in the observableuniverse

the designer have not time to create table

no agent could ever learn the right table entries from its experience

how to fill in table entries?

29

Page 30: Intelligent Agents - Sharif

Agent program

Mapping is not necessarily using a table.

AI intends to find programs producing rational behavior (to the

extent possible) from a smallish program instead of a vast table

Can AI do for general intelligent behavior what Newton did for

square roots?

function SQRT( double X )

{

double r = 1.0 ;

while ( fabs( r * r - x ) > 10-8 )

r = r - ( r * r - x ) / 2r ;

return r ;

}

1.0 1.00000

1.1 1.04898

1.2 1.09565

1.3 1.14056

...

Huge tables of square roots before 1970s

30

Page 31: Intelligent Agents - Sharif

Simple Reflex Agents

Agent Program

31

Page 32: Intelligent Agents - Sharif

Simple Reflex Agents

Select actions on the basis of the current percept

ignoring the rest of the percept history

If car-in-front-is-braking then initiate-braking

Blinking reflex

32

Page 33: Intelligent Agents - Sharif

Simple Reflex Agents

Simple, but very limited intelligence

Works only if the correct decision can be made on the

basis of the current percept (fully observability)

Infinite loops in partially observable environment

33

Page 34: Intelligent Agents - Sharif

Model-based reflex agents

34

Page 35: Intelligent Agents - Sharif

Model-based reflex agents

Partial observability

Internal state (based on percept history)

reflects some unobserved aspects of the current state

Updating the internal state information requires two kinds of

knowledge

Information about how the world evolves (independent of agent)

Information about how the agent's own actions affects the world

Only determine the best guess for the current state of a

partially observable environment

35

Page 36: Intelligent Agents - Sharif

Goal-based agents

36

Page 37: Intelligent Agents - Sharif

Goal-based agents

Knowing about the current state is not always enough to

decide what to do

Situations that are desirable must be specified (goal)

Usually requires search and planning

to find action sequences achieving goal

37

Page 38: Intelligent Agents - Sharif

Goal-based agents vs. reflex-based agents

Consideration of future

Goal-based agents may be less efficient but are more

flexible

Knowledge is represented explicitly and can be changed easily

Example: going to a new destination

Goal-based agent: specifying that destination as the goal

Reflexive agent: agent's rules for when to turn and when to go straight

must be rewritten

38

Page 39: Intelligent Agents - Sharif

Utility-based agents

39

Page 40: Intelligent Agents - Sharif

Utility-based agents

More general performance measure than goals How happy would each world state make the agent?

Utility function is an internalization of performance measure

Advantages Like goal-based agents show flexibility and learning advantages

Can trade-off conflicting goals (e.g. speed and safety)

Where none of several goals can be achieved with certainty

likelihood of success can be weighted by importance of goals

Rational utility-based agent chooses the action that maximizesthe expected utility of action outcomes Many chapters of AIMA book is about this

Handling uncertainty in partially observable and stochastic environments

40

Page 41: Intelligent Agents - Sharif

Learning Agents

41

Page 42: Intelligent Agents - Sharif

Learning Agents

Create state-of-the-art systems in many areas of AI

Four conceptual components Performance element: selects actions based on percepts (considered as

entire agent before)

Learning element: makes improvements by modifying “knowledge”(performance element) based on critic feedback

Critic: feedbacks on how the agents is doing

Problem generator: suggests actions leading to new and informativeexperiences

Performance standard Fixed, out of the agent

Percepts themselves do not provide indication of success

Distinguishes part of the incoming percept as a reward or penalty

42

Page 43: Intelligent Agents - Sharif

Learning Agents

Learning element is based on performance element (i.e.,

agent design)

The learning element can make changes to any of the

"knowledge" components of previous agent diagrams

To bring components into closer agreement with feedback

(yielding better overall performance)

43