Decision-Theoretic Military Operations Planning

Military operations planning involves concurrent actions, resource assignment, and conflicting costs. Individual tasks sometimes fail with a known probability, promoting a decision-theoretic approach. The planner must choose between multiple tasks that achieve similar outcomes but have different costs. The military domain is particularly suited to automated methods because hundreds of tasks, specified by many planning staff, need to be quickly and robustly coordinated. The authors are not aware of any previous planners that handle all characteristics of the operations planning domain in a single package. This paper shows that problems with such features can be successfully approached by realtime heuristic search algorithms, operating on a formulation of the problem as a Markov decision process. Novel automatically generated heuristics, and classic caching methods, allow problems of interesting sizes to be handled. Results are presented on data provided by the Australian Defence Science and Technology Organisation.

[1]  R. Bellman,et al.  Dynamic Programming and Markov Processes. , 1961 .

[2]  Austin Tate Advanced Planning Technology: Technological Achievements of the ARPA/Rome Laboratory Planning Inititive , 1996 .

[3]  Austin Tate,et al.  O-Plan: a Knowledge-Based Planner and its Application to Logistics , 1996 .

[4]  Kee-Eung Kim,et al.  Solving Very Large Weakly Coupled Markov Decision Processes , 1998, AAAI/IAAI.

[5]  Jim Blythe,et al.  Decision-Theoretic Planning , 1999, AI Mag..

[6]  Alice M. Mulvehill,et al.  JADE: A Tool for Rapid Crisis Action Planning , 1999 .

[7]  John Langford,et al.  Probabilistic Planning in the Graphplan Framework , 1999, ECP.

[8]  Edmund H. Durfee,et al.  Plan generation and hard real-time execution with application to safe, autonomous flight , 1999 .

[9]  Jesse Hoey,et al.  SPUDD: Stochastic Planning using Decision Diagrams , 1999, UAI.

[10]  Shlomo Zilberstein,et al.  LAO*: A heuristic search algorithm that finds solutions with loops , 2001, Artif. Intell..

[11]  Fahiem Bacchus,et al.  Planning with Resources and Concurrency: A Forward Chaining Approach , 2001, IJCAI.

[12]  Hector Geffner,et al.  Heuristic Planning with Time and Resources , 2001 .

[13]  Lex Weaver,et al.  A Multi-Agent Policy-Gradient Approach to Network Routing , 2001, ICML.

[14]  Peter Geibel Reinforcement Learning with Bounded Risk , 2001, ICML.

[15]  Sanjoy Dasgupta,et al.  Off-Policy Temporal Difference Learning with Function Approximation , 2001, ICML.

[16]  Zhengzhu Feng,et al.  Symbolic LAO* Search for Factored Markov Decision Processes , 2002, AAAI 2002.

[17]  Jonathan Billington,et al.  A coloured petri net based tool for course of action development and analysis , 2002, FME 2002.

[18]  Ivan Serina,et al.  LPG: A Planner Based on Local Search for Planning Graphs with Action Costs , 2002, AIPS.

[19]  Subbarao Kambhampati,et al.  Planning Graph-based Heuristics for Cost-sensitive Temporal Planning , 2002, AIPS.

[20]  David J. Musliner,et al.  A Framework for Planning in Continuous-time Stochastic Domains , 2003, ICAPS.

[21]  Blai Bonet,et al.  Labeled RTDP: Improving the Convergence of Real-Time Dynamic Programming , 2003, ICAPS.

[22]  Kathryn B. Laskey,et al.  Limited Rationality in Action: Decision Support for Military Situation Assessment , 2004, Minds and Machines.

[23]  H. Thuve TOPFAS ( Tool for Operational Planning , Force Activation and Simulation ) , .