Online search agents operate in unknown environments and must learn through interaction rather than pure computation. They interleave planning and action by first taking an action, observing the outcome, and using this information to plan their next action. Like hill-climbing search, online search agents keep only their current state in memory, acting locally to explore their environment through experimentation in real-time.
Online search agents operate in unknown environments and must learn through interaction rather than pure computation. They interleave planning and action by first taking an action, observing the outcome, and using this information to plan their next action. Like hill-climbing search, online search agents keep only their current state in memory, acting locally to explore their environment through experimentation in real-time.
Online search agents operate in unknown environments and must learn through interaction rather than pure computation. They interleave planning and action by first taking an action, observing the outcome, and using this information to plan their next action. Like hill-climbing search, online search agents keep only their current state in memory, acting locally to explore their environment through experimentation in real-time.
Online search agents operate in unknown environments and must learn through interaction rather than pure computation. They interleave planning and action by first taking an action, observing the outcome, and using this information to plan their next action. Like hill-climbing search, online search agents keep only their current state in memory, acting locally to explore their environment through experimentation in real-time.
Download as PPTX, PDF, TXT or read online from Scribd
Download as pptx, pdf, or txt
You are on page 1of 13
ONLINE SEARCH AGENTS
• offline search algorithms -They compute a complete solution
before setting foot. in the real world and then execute the solution • Online search agent interleaves computation and action: first it takes an action, then it observes the environment and computes the next action • Online search is a necessary idea for unknown environments, where the agent does not know what states exist or what its actions do. • In this state of ignorance, the agent faces an exploration problem and must use its actions as experiments in order to learn enough to make deliberation worthwhile. • The baby‘s gradual discovery of how the world works is, in part, an online search process. Online search problems • An online search problem must be solved by an agent executing actions, rather than by pure computation • the agent knows only the following: – ACTIONS(s), which returns a list of actions allowed in state s;
– The step-cost function c( s, a, s')-note that this cannot be used until
the agent knows that s' is the outcome; and – GOAL-TEST(s). Online search agents • After each action, an online agent receives a percept telling it what state it has reached; from this information, it can augment its map of the environment. • The current map is used to decide where to go next. • This interleaving of planning and action means that online search algorithms are quite different from the offline search algorithms Online local search • Like depth-first search, hill-climbing search has the property of locality in its node expansions. • In fact, because it keeps just one current state in memory, hill- climbing search is already an online search algorithm • Instead of random restarts, one might consider using a random walk to explore the environment. • A random walk simply selects at random one of the available actions from the current state • preference can be given to actions that have not yet been tried. It is easy to prove that a random walk will eventually find a goal or complete its exploration, provided that the space is finite • Learning Real time A *. An LRTA* agent is guaranteed to find a goal in any finite, safely explorable environment.