Planning Under Continuous Time and Resource Uncertainty: A Challenge for AI

We outline a class of problems, typical of Mars rover operations, that are problematic for current methods of planning under uncertainty. The existing methods fail because they suffer from one or more of the following limitations: 1) they rely on very simple models of actions and time, 2) they assume that uncertainty is manifested in discrete action outcomes, 3) they are only practical for very small problems. For many real world oroblems, these assumptions fail to hold. In particular, when planning the activities for a Mars rover, none of the above assumptions is valid: 1) actions can be concurrent and have differing durations, 2) there is uncertainty concerning action durations and consumption of continuous resources like power, and 3) typical daily plans involve on the order of a hundred actions. This class of problems may be of particular interest to the UAI community because both classical and decision-theoretic planning techniques may be useful in solving it. We describe the rover problem, discuss previous work on planning under uncertainty, and present a detailed, but very small, example illustrating some of the difficulties of finding good plans.

[1]  Keith Golden,et al.  Leap Before You Look: Information Gathering in the PUCCINI Planner , 1998, AIPS.

[2]  Michael L. Littman,et al.  Contingent planning under uncertainty via stochastic satisfiability , 1999, Artif. Intell..

[3]  Oren Etzioni,et al.  An Approach to Planning with Incomplete Information , 1992, KR.

[4]  John L. Bresina,et al.  Just-In-Case Scheduling , 1994, AAAI.

[5]  David E. Smith,et al.  Extending Graphplan to Handle Uncertainty & Sensing Actions , 1998, AAAI/IAAI.

[6]  Daniel S. Weld,et al.  Probabilistic Planning with Information Gathering and Contingent Execution , 1994, AIPS.

[7]  Robert P. Goldman,et al.  Conditional Linear Planning , 1994, AIPS.

[8]  Jussi Rintanen,et al.  Constructing Conditional Plans by a Theorem-Prover , 1999, J. Artif. Intell. Res..

[9]  Blai Bonet,et al.  Planning with Incomplete Information as Heuristic Search in Belief Space , 2000, AIPS.

[10]  Mark A. Peot,et al.  Conditional nonlinear planning , 1992 .

[11]  Martin L. Puterman,et al.  Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .

[12]  Enrico Giunchiglia,et al.  Planning as Satisfiability in Nondeterministic Domains , 2000, AAAI/IAAI.

[13]  Satinder P. Singh,et al.  How to Dynamically Merge Markov Decision Processes , 1997, NIPS.

[14]  Leslie Pack Kaelbling,et al.  Practical Reinforcement Learning in Continuous Spaces , 2000, ICML.

[15]  David E. Smith,et al.  Conformant Graphplan , 1998, AAAI/IAAI.

[16]  David H. D. Warren,et al.  Generating Conditional Plans and Programs , 1976, AISB.

[17]  Amos Storkey,et al.  Advances in Neural Information Processing Systems 20 , 2007 .

[18]  Jim Blythe,et al.  Decision-Theoretic Planning , 1999, AI Mag..

[19]  Marco Roveri,et al.  Conformant Planning via Symbolic Model Checking , 2000, J. Artif. Intell. Res..

[20]  Rémi Munos,et al.  A Study of Reinforcement Learning in the Continuous Case by the Means of Viscosity Solutions , 2000, Machine Learning.

[21]  Ross D. Shachter,et al.  Decision-theoretic planning , 1998 .

[22]  Michael L. Littman,et al.  Exact Solutions to Time-Dependent MDPs , 2000, NIPS.

[23]  John Langford,et al.  Probabilistic Planning in the Graphplan Framework , 1999, ECP.

[24]  David E. Smith,et al.  Extending Graphplan to handle uncertainty and sensing actions , 1998, AAAI 1998.

[25]  Andrew W. Moore,et al.  Variable Resolution Discretization for High-Accuracy Solutions of Optimal Control Problems , 1999, IJCAI.

[26]  Piergiorgio Bertoli,et al.  Heuristic Search + Symbolic Model Checking = Efficient Conformant Planning , 2001, IJCAI.

[27]  Piergiorgio Bertoli,et al.  Planning in Nondeterministic Domains under Partial Observability via Symbolic Model Checking , 2001, IJCAI.

[28]  Michael L. Littman,et al.  The Computational Complexity of Probabilistic Planning , 1998, J. Artif. Intell. Res..

[29]  Sebastian Thrun,et al.  Monte Carlo POMDPs , 1999, NIPS.

[30]  Jim Blythe,et al.  Planning Under Uncertainty in Dynamic Domains , 1998 .

[31]  Kee-Eung Kim,et al.  Solving Very Large Weakly Coupled Markov Decision Processes , 1998, AAAI/IAAI.

[32]  Gregg Collins,et al.  Planning for Contingencies: A Decision-based Approach , 1996, J. Artif. Intell. Res..

[33]  P. Pandurang Nayak,et al.  Fragment-based Conformant Planning , 2002, AIPS.

[34]  Craig Boutilier,et al.  Decision-Theoretic Planning: Structural Assumptions and Computational Leverage , 1999, J. Artif. Intell. Res..

[35]  Nicholas Kushmerick,et al.  An Algorithm for Probabilistic Planning , 1995, Artif. Intell..

[36]  John N. Tsitsiklis,et al.  Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.

[37]  Martha E. Pollack,et al.  Conditional, Probabilistic Planning: A Unifying Algorithm and Effective Search Control Mechanisms , 1999, AAAI/IAAI.