《认知机器人》(英文版) Massachusetts Institute of Technology

Massachusetts Institute of Technology 16.412/6.834 Cognitive Robotics Distributed: Monday, 3/31/04 objective The purpose of the following handout is to walk you step by step through the execution of the ff planning algorithm, on a simple example. The ff algorithm is presented in the The FF Planning System: Fast Plan Generation Through Heuristic Search, by Jorg Hoffmann and Bernhard Nebel, Journal of Artificial Intelligence Research, 2001 Problem by a door), with the Polish operator added in move betveen rooms A and B, separated You are given operators for the door dome Door-2 Domain Operators Move(x, y) Pre: in(x) Add: in(y) Del Pre: closed Add: opened Del closed Close Pre Polish Pre: opened Add: polished Del: Consider the problem of generating a plan that moves from the initial state In(A) Closed
Massachusetts Institute of Technology 16.412/6.834 Cognitive Robotics Distributed: Monday, 3/31/04 Objective The purpose of the following handout is to walk you step by step through the execution of the FF planning algorithm, on a simple example. The FF algorithm is presented in the paper: The FF Planning System: Fast Plan Generation Through Heuristic Search,” by Jorg Hoffmann and Bernhard Nebel, Journal of Artificial Intelligence Research, 2001. Problem You are given operators for the door domain (move between rooms A and B, separated by a door), with the Polish operator added. Door-2 Domain Operators Move (x,y) Pre: in(x) opened Add: in(y) Del: in(x) Open Pre: closed Add: opened Del: closed Close Pre: opened Add: closed Del: opened Polish Pre: opened Add: polished Del: Consider the problem of generating a plan that moves from the initial state: In(A) Closed

To the goal state In(B) In the following we develop the tree of nodes visited in the state space search by the FF algorithm. We label each node with its relaxed distance estimate. For each node we demonstrate how the relaxed distance estimate is computed, including the plan grap generated from the node to the goal state, the relaxed plan that is extracted, and the resulting heuristic cost of that node FF starts with the initial state as the root of the tree Node 1 In(A) Estimate? Closed FF then computes the heuristic value of Node 1 by generating a relaxed plan graph, from Node I to the goal In(a) noop In(a) nooD In(a) Move(A B In(B) Closed noo Closed Closed Opened noop Opened Polish Each operator is relaxed by simply eliminating its delete list. The relaxed plan graph is similar to that produced by GraphPlan, except that it does not contain any mutual exclusion relations. The relaxed plan graph for Node 1 consists of three fact layers and two action layers. The three goal propositions appear in the third fact layer Note that, for simplicity, in our drawing we have drawn each action once, in the action layer in which it first appears. That action should also appear in all subsequent layers However, its consequences are also achieved through null operations(noops). For example, in the graph above, FF or Graph Plan would include an arc from Opened in layer 2 to Closed in layer 3, representing the Close action. When generating the relaxed plan, FF never selects these subsequent actions, since it al ways prefers noops to actions Hence, they serve no useful role
To the goal state: In(B) Closed Polished In the following we develop the tree of nodes visited in the state space search by the FF algorithm. We Label each node with its relaxed distance estimate. For each node we demonstrate how the relaxed distance estimate is computed, including the plan graph generated from the node to the goal state, the relaxed plan that is extracted, and the resulting heuristic cost of that node. FF starts with the initial state as the root of the tree: FF then computes the heuristic value of Node 1 by generating a relaxed plan graph, from Node 1 to the goal: Each operator is relaxed by simply eliminating its delete list. The relaxed plan graph is similar to that produced by GraphPlan, except that it does not contain any mutual exclusion relations. The relaxed plan graph for Node 1 consists of three fact layers and two action layers. The three goal propositions appear in the third fact layer. Note that, for simplicity, in our drawing we have drawn each action once, in the action layer in which it first appears. That action should also appear in all subsequent layers. However, its consequences are also achieved through null operations (noops). For example, in the graph above, FF or GraphPlan would include an arc from Opened in layer 2 to Closed in layer 3, representing the Close action. When generating the relaxed plan, FF never selects these subsequent actions, since it always prefers noops to actions. Hence, they serve no useful role. In(A) Closed In(A) Closed Opened In(A) In(B) Closed Opened Polished noop noop Open noop noop noop Polish Move(A,B) Node 1 In(A) Estimate ? Closed Close

To generate the relaxed plan, FF starts with the three goal propositions In(B), closed, Polished For each goal it selects a noop or action from action layer 2 that achieves it, preferring no-ops over actions. The selected actions/noops are highlighted in bold. In (B)is achieved by action Move(A, B), Closed is carried forward by a noop, and Polished is achieved using the Polish action. The corresponding subgoals for fact layer 2 are collected and consist of In(A), Closed, Open Next, actions and noops from action layer l are selected to achieve these subgoals. In(A) and Closed are achieved by noops, while Open is achieved with the Close action. The corresponding subgoals are all propositions in the initial state relaxed plan, shown in bold, consists of three actions--Open, Move(A, B)and polish, a The heuristic value of Node 1 is the number of actions in the relaxed plan. The comple lence. the heuristic value of node 1 is 3 FF expands Node 1 by considering the application of each action in the first layer of the relaxed plan. Note that FF only uses an action in this first layer if the effects of the action are used by an action in layer 2. ff stops evaluating actions as soon as it finds an action whose target state has a better heuristic value. If the heuristic value does not improve, it performs a breadth first search For Node 1, only one action, Closed, appears in the first layer of the relaxed plan, hence it is the only action expanded. The expanded graph is shown below, with Node 2 added Node 1 In(A) Estimate 3 Closed Node 2 In(A)Estimate? Opened lext we generate the relaxed plan graph and plan for Node 2, as shown below
To generate the relaxed plan, FF starts with the three goal propositions: In(B), Closed, Polished. For each goal it selects a noop or action from action layer 2 that achieves it, preferring no-ops over actions. The selected actions/noops are highlighted in bold. In(B) is achieved by action Move(A,B), Closed is carried forward by a noop, and Polished is achieved using the Polish action. The corresponding subgoals for fact layer 2 are collected, and consist of: In(A), Closed, Open Next, actions and noops from action layer 1 are selected to achieve these subgoals. In(A) and Closed are achieved by noops, while Open is achieved with the Close action. The corresponding subgoals are all propositions in the initial state. The heuristic value of Node 1 is the number of actions in the relaxed plan. The complete relaxed plan, shown in bold, consists of three actions -- Open, Move(A,B) and Polish – hence, the heuristic value of Node 1 is 3. FF expands Node 1 by considering the application of each action in the first layer of the relaxed plan. Note that FF only uses an action in this first layer if the effects of the action are used by an action in layer 2. FF stops evaluating actions as soon as it finds an action whose target state has a better heuristic value. If the heuristic value does not improve, it performs a breadth first search. For Node 1, only one action, Closed, appears in the first layer of the relaxed plan, hence it is the only action expanded. The expanded graph is shown below, with Node 2 added: Next we generate the relaxed plan graph and plan for Node 2, as shown below: Node 1 In(A) Estimate 3 Closed Estimate ? Open Node 2 In(A) Opened

In(a n(B) noop Opened Polish Polished The plan graph has one action layer, separating the current state and goal state fact layers The relaxed plan consists of three parallel actions: Move(A, B), Close, and Polish, resulting in a cost estimate of 3. The node value is the same as for Node 1 indicating a plateau. Hence, actions are explored in breadth first order until the first action is found with decreased heuristic cost. In this example, we evaluate each of the three concurrent actions of Node 2s plan, working in a top-down order, as they are written in the relaxed plan. However, no specific order for expanding the breadth first search is required First we expand the action Move(A, B) Node 1 In(a) Estimate 3 Closed Open Node 2 In(a) Estimate 3 Move(A, B) Node 3 In(B)Estimate? Opened Next we calculate the relaxed plan graph estimate for Node 3
In(A) Opened In(A) In(B) Closed Opened Polished noop noop Close Move(A,B) Polish The plan graph has one action layer, separating the current state and goal state fact layers. The relaxed plan consists of three parallel actions: Move(A,B), Close, and Polish, resulting in a cost estimate of 3. The node value is the same as for Node 1, indicating a plateau. Hence, actions are explored in breadth first order until the first action is found with decreased heuristic cost. In this example, we evaluate each of the three concurrent actions of Node 2’s plan, working in a top-down order, as they are written in the relaxed plan. However, no specific order for expanding the breadth first search is required. First we expand the action Move(A,B): Next we calculate the relaxed plan graph estimate for Node 3: Node 1 In(A) Estimate 3 Closed Estimate 3 Open Node 2 In(A) Opened Estimate ? Move(A,B) Node 3 In(B) Opened

In(B) n(B) ose Opened Once again the plan graph has a single action layer. The relaxed plan only contains two actions: Close and Polish, hence the cost of Node 3 is 2. This estimate is strictly better than that of Nodes l and 2; hence Node 3 moves FF off the plateau, and is chosen for further expansion. FF has two actions available to expand Node 3, either Close or Polish First choosing Close Node 4 is added to the search tree n(A) Estimate 3 n(A)Estimate Opened Move(A, B) In(B) Estimate 2 Opened Clos Polish Node 4 In(B) Estimate? The relaxed plan graph and plan for Node 4 is as follows
In(A) In(B) Closed Opened Polished In(B) Opened noop noop Close Move Polish Once again the plan graph has a single action layer. The relaxed plan only contains two actions: Close and Polish, hence the cost of Node 3 is 2. This estimate is strictly better than that of Nodes 1 and 2; hence Node 3 moves FF off the plateau, and is chosen for further expansion. FF has two actions available to expand Node 3, either Close or Polish. First, choosing Close, Node 4 is added to the search tree: Node 1 In(A) Estimate 3 Closed Open The relaxed plan graph and plan for Node 4 is as follows: Node 2 In(A) Estimate 3 Opened Move(A,B) Node 3 In(B) Estimate 2 Opened Estimate ? Close Polish Node 4 In(B) Closed

Move In(a) In(B) noo In(B) In(B) Closed Closed Closed Opened Polished Note that this plan graph has two action layers, one to open the door, and the second to act based on the open door (also note that the plan graph includes move operations, but are not indicated here for simplicity). The relaxed plan consists of Open, followed by Polish, with a heuristic cost of 2. This does not represent an improvement over Node 3, ence we try expanding Node 3 using action Polish, producing Node 5 Node 1 In(a) Estimate 3 Closed Open Node 2 In(a) Estimate 3 Opened Move(A B Polish Node 3 In(B) Estimate 2 Opened Polish Node 4 n(B)Estimate 2 In(B)Estimate? Polished The relaxed plan graph for Node 5 is
Polish noop noop noop Close Move In(A) In(B) Closed Opened Polished In(B) Closed Opened In(B) Closed noop Open noop Note that this plan graph has two action layers, one to open the door, and the second to act based on the open door (also note that the plan graph includes move operations, but are not indicated here for simplicity). The relaxed plan consists of Open, followed by Polish, with a heuristic cost of 2. This does not represent an improvement over Node 3, hence we try expanding Node 3 using action Polish, producing Node 5: Node 1 In(A) Estimate 3 Closed Open The relaxed plan graph for Node 5 is: Node 2 In(A) Estimate 3 Opened Move(A,B) Polish Node 3 In(B) Estimate 2 Opened Estimate 2 Close In(B) Closed Node 4 In(B) Estimate ? Opened Polished Polish Node 5

ove In(B) n(B) ose Opened Polished Polished It has a single action, Close hence the estimated cost for Node 5 is 1. This is strictly better than the cost at Node 3, so FF continues its expansion from Node 5. One action is available for expansion, Close, resulting in Node 6 Node 1 In(A)Estimate 3 Closed Open Node 2 n(A)Estimate 3 Opened Polish Move(A, B Node 3 In(B) Estimate 2 Opened Close Polish Node 4 In(B) Estimate 2 Node 5 In(B)Estimate 1 Polished Close Node 6 In(B)Estimate Polished
It has a single action, Close, hence the estimated cost for Node 5 is 1. This is strictly better than the cost at Node 3, so FF continues its expansion from Node 5. One action is available for expansion, Close, resulting in Node 6: In(A) Estimate 3 Closed Node 1 In(A) Estimate 3 Opened Node 2 Open In(B) Estimate 2 Opened Node 3 Move(A,B) In(B) Estimate 2 Closed Node 4 Close In(B) Estimate 1 Opened Polished Node 5 Polish In(B) Estimate ? Closed Polished Node 6 Close In(B) Opened Polished In(A) In(B) noop Closed Opened Polished noop Close Move noop Polish

FF finds that all goal propositions are achieved at Node 6, hence the plan graph contains a single fact layer, and has an estimated cost of 0 Polished The completed plan is the action sequence from Node 1 to 6, consisting of Open Move(A, B), polish, Close Node 1 In(a) Estimate 3 Closed Node 2 In(a) Estimate 3 Move(A, B) Node 3 In(B) Estimate 2 Close Polish Node 4 In(B)Estimate 2 Node 5 In(B) Estimate 1 Closed Opened Polished Close Node 6 In(B) Estimate O Closed Polished
FF finds that all goal propositions are achieved at Node 6, hence the plan graph contains a single fact layer, and has an estimated cost of 0: The completed plan is the action sequence from Node 1 to 6, consisting of Open, Move(A,B), Polish, Close: In(B) Closed Polished Node 1 In(A) Estimate 3 Closed Open Node 2 In(A) Estimate 3 Opened Estimate 2 Move(A,B) Polish Node 3 In(B) Opened Estimate 2 Close In(B) Closed Node 4 In(B) Estimate 1 Opened Polished Polish Node 5 Estimate 0 Close Node 6 In(B) Closed Polished
按次数下载不扣除下载券;
注册用户24小时内重复下载只扣除一次;
顺序:VIP每日次数-->可用次数-->下载券;
- 《认知机器人》(英文版) Distributed constraint Satisfaction problems.pdf
- 《认知机器人》(英文版) LPG: Local search for Planning Graphs.pdf
- 《认知机器人》(英文版) Fast Solutions to CSPs.pdf
- 《认知机器人》(英文版) Planning as Heuristic Forward Search.pdf
- 《认知机器人》(英文版) Using the Forest to See the Trees Context-based Object Recognition.pdf
- 《认知机器人》(英文版) Hybrid Mode Estimation and Gaussian Filtering with Hybrid HMMs.pdf
- 《认知机器人》(英文版) Model-based Programming and Constraint-based HMMs.pdf
- 《认知机器人》(英文版) Optimal csPs and Conflict-directed.pdf
- 《认知机器人》(英文版) Mapping Topics: Topological Maps.pdf
- 《认知机器人》(英文版) Conflict-directed Diagnosis and Probabilistic Mode Estimation.pdf
- 《认知机器人》(英文版) Fault Aware Systems: Model-based Programming and Diagnosis.pdf
- 《认知机器人》(英文版) Foundations of state Estimation.pdf
- 《认知机器人》(英文版) Temporal Planning in Space.pdf
- 《认知机器人》(英文版) Executing Model-based Programs Using.pdf
- 《认知机器人》(英文版) Foundations of State Estimation PartⅡ.pdf
- 《认知机器人》(英文版) Robot Motion Planning and (a little)Computational Geometry.pdf
- 《认知机器人》(英文版) Probabilistic methods for Kinodynamic Path Planning.pdf
- 《认知机器人》(英文版) Incremental Path Planning.pdf
- 《认知机器人》(英文版) Path Planning in Partially-Known Environments.pdf
- 《认知机器人》(英文版) Course Objective 1.pdf
- 《认知机器人》(英文版) Fast Solutions to CSp's.pdf
- 《认知机器人》(英文版) Reactive Planning in Large State Spaces Through.pdf
- 《认知机器人》(英文版) Partially Observable Markov Decision Processes.pdf
- 《认知机器人》(英文版) Vision-based SLAM.pdf
- 《认知机器人》(英文版) SIFT SLAM Vision Details.pdf
- 《认知机器人》(英文版) Information Based Adaptive Robotic Exploration.pdf
- 《认知机器人》(英文版) Temporal Plan Execution: Dynamic Scheduling and Simple Temporal Networks.pdf
- 《认知机器人》(英文版) Partially Observable Markov Decision Processes Part II.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义,英文版)Handout 1:Bode plot rules reminder.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义,英文版)Handout 2:Gain and Phase margins.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义,英文版)Handout 5:Control System Design Principles.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义,英文版)Handout 6:Proportional Compensation.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)Handout 3:Gain and Phase Margins for unstable.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)Handout 4:Root-Locus Review.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)extra.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)Handout 7:Lag and PI compensation.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)Handout 8:Lead compensation.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)Handout 8:Lead compensation.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)Handout 12:Plants with right half-plane zeros.pdf
- 美国麻省理工大学:《航空系统的估计与控制》教学资源(讲义)Handout 10:Notch compensation.pdf