TY - GEN
T1 - Point-based policy transformation
T2 - 10th International Workshop on the Algorithmic Foundations of Robotics, WAFR 2012
AU - Kurniawati, Hanna
AU - Patrikalakis, Nicholas M.
N1 - Publisher Copyright:
© Springer-Verlag Berlin Heidelberg 2013.
PY - 2013
Y1 - 2013
N2 - Motion planning under uncertainty that can efficiently take into account changes in the environment is critical for robots to operate reliably in our living spaces. Partially Observable Markov Decision Process (POMDP) provides a systematic and general framework for motion planning under uncertainty. Point-based POMDP has advanced POMDP planning tremendously over the past few years, enabling POMDP planning to be practical for many simple to moderately difficult robotics problems. However, when environmental changes alter the POMDP model, most existing POMDP planners recompute the solution from scratch, often wasting significant computational resources that have been spent for solving the original problem. In this paper, we propose a novel algorithm, called Point-Based Policy Transformation (PBPT), that solves the altered POMDP problem by transforming the solution of the original problem to accommodate changes in the problem. PBPT uses the point-based POMDP approach. It transforms the original solution by modifying the set of sampled beliefs that represents the belief space B, and then uses this new set of sampled beliefs to revise the original solution. Preliminary results indicate that PBPT generates a good policy for the altered POMDP model in a matter of minutes, while recomputing the policy using the fastest offline POMDP planner today fails to find a policy with similar quality after two hours of planning time, even when the policy for the original problem is reused as an initial policy.
AB - Motion planning under uncertainty that can efficiently take into account changes in the environment is critical for robots to operate reliably in our living spaces. Partially Observable Markov Decision Process (POMDP) provides a systematic and general framework for motion planning under uncertainty. Point-based POMDP has advanced POMDP planning tremendously over the past few years, enabling POMDP planning to be practical for many simple to moderately difficult robotics problems. However, when environmental changes alter the POMDP model, most existing POMDP planners recompute the solution from scratch, often wasting significant computational resources that have been spent for solving the original problem. In this paper, we propose a novel algorithm, called Point-Based Policy Transformation (PBPT), that solves the altered POMDP problem by transforming the solution of the original problem to accommodate changes in the problem. PBPT uses the point-based POMDP approach. It transforms the original solution by modifying the set of sampled beliefs that represents the belief space B, and then uses this new set of sampled beliefs to revise the original solution. Preliminary results indicate that PBPT generates a good policy for the altered POMDP model in a matter of minutes, while recomputing the policy using the fastest offline POMDP planner today fails to find a policy with similar quality after two hours of planning time, even when the policy for the original problem is reused as an initial policy.
UR - http://www.scopus.com/inward/record.url?scp=85009517018&partnerID=8YFLogxK
U2 - 10.1007/978-3-642-36279-8_30
DO - 10.1007/978-3-642-36279-8_30
M3 - Conference contribution
AN - SCOPUS:85009517018
SN - 9783642362781
T3 - Springer Tracts in Advanced Robotics
SP - 493
EP - 509
BT - Springer Tracts in Advanced Robotics
A2 - Frazzoli, Emilio
A2 - Roy, Nicholas
A2 - Lozano-Perez, Tomas
A2 - Rus, Daniela
PB - Springer Verlag Italia
Y2 - 13 June 2012 through 15 June 2012
ER -