Dynamic Programming for POMDPs Using a Factored State Representation
暂无分享,去创建一个
[1] Chelsea C. White,et al. A survey of solution techniques for the partially observed Markov decision process , 1991, Ann. Oper. Res..
[2] Craig Boutilier,et al. Exploiting Structure in Policy Construction , 1995, IJCAI.
[3] Daniel S. Weld,et al. Probabilistic Planning with Information Gathering and Contingent Execution , 1994, AIPS.
[4] Michael L. Littman,et al. Incremental Pruning: A Simple, Fast, Exact Method for Partially Observable Markov Decision Processes , 1997, UAI.
[5] M. Littman. The Witness Algorithm: Solving Partially Observable Markov Decision Processes , 1994 .
[6] R. I. Bahar,et al. Algebraic decision diagrams and their applications , 1993, Proceedings of 1993 International Conference on Computer Aided Design (ICCAD).
[7] Martha E. Pollack,et al. Conditional, Probabilistic Planning: A Unifying Algorithm and Effective Search Control Mechanisms , 1999, AAAI/IAAI.
[8] Enrico Macii,et al. Algebraic decision diagrams and their applications , 1993, Proceedings of 1993 International Conference on Computer Aided Design (ICCAD).
[9] Eric A. Hansen,et al. Solving POMDPs by Searching in Policy Space , 1998, UAI.
[10] Craig Boutilier,et al. Decision-Theoretic Planning: Structural Assumptions and Computational Leverage , 1999, J. Artif. Intell. Res..
[11] J SondikEdward. The Optimal Control of Partially Observable Markov Processes over the Infinite Horizon , 1978 .
[12] Edward J. Sondik,et al. The Optimal Control of Partially Observable Markov Processes over the Infinite Horizon: Discounted Costs , 1978, Oper. Res..
[13] Craig Boutilier,et al. Computing Optimal Policies for Partially Observable Decision Processes Using Compact Representations , 1996, AAAI/IAAI, Vol. 2.
[14] Jesse Hoey,et al. SPUDD: Stochastic Planning using Decision Diagrams , 1999, UAI.
[15] Edward J. Sondik,et al. The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..