Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Online Search Agents

Download as pptx, pdf, or txt
Download as pptx, pdf, or txt
You are on page 1of 13

ONLINE SEARCH AGENTS

• offline search algorithms -They compute a complete solution


before setting foot. in the real world and then execute the
solution
• Online search agent interleaves computation and action: first
it takes an action, then it observes the environment and
computes the next action
• Online search is a necessary idea for unknown environments,
where the agent does not know what states exist or what its
actions do.
• In this state of ignorance, the agent faces an exploration
problem and must use its actions as experiments in order to
learn enough to make deliberation worthwhile.
• The baby‘s gradual discovery of how the world works is, in
part, an online search process.
Online search problems
• An online search problem must be solved by an agent
executing actions, rather than by pure computation
• the agent knows only the following:
– ACTIONS(s), which returns a list of actions allowed in state s;

– The step-cost function c( s, a, s')-note that this cannot be used until


the agent knows that s' is the outcome; and
– GOAL-TEST(s).
Online search agents
• After each action, an online agent receives a percept telling it
what state it has reached; from this information, it can
augment its map of the environment.
• The current map is used to decide where to go next.
• This interleaving of planning and action means that online
search algorithms are quite different from the offline search
algorithms
Online local search
• Like depth-first search, hill-climbing search has the property
of locality in its node expansions.
• In fact, because it keeps just one current state in memory, hill-
climbing search is already an online search algorithm
• Instead of random restarts, one might consider using a
random walk to explore the environment.
• A random walk simply selects at random one of the available
actions from the current state
• preference can be given to actions that have not yet been
tried. It is easy to prove that a random walk will eventually
find a goal or complete its exploration, provided that the
space is finite
• Learning Real time A *. An LRTA* agent is guaranteed to find a
goal in any finite, safely explorable environment.

You might also like