Speeding Up the Convergence of Value Iteration in Partially Observable Markov Decision Processes
暂无分享,去创建一个
[1] Karl Johan Åström,et al. Optimal control of Markov processes with incomplete state information , 1965 .
[2] Edward J. Sondik,et al. The optimal control of par-tially observable Markov processes , 1971 .
[3] E. J. Sondik,et al. The Optimal Control of Partially Observable Markov Decision Processes. , 1971 .
[4] Edward J. Sondik,et al. The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..
[5] Drew McDermott,et al. Planning and Acting , 1978, Cogn. Sci..
[6] Edward J. Sondik,et al. The Optimal Control of Partially Observable Markov Processes over the Infinite Horizon: Discounted Costs , 1978, Oper. Res..
[7] James N. Eagle. The Optimal Search for a Moving Target When the Search Path Is Constrained , 1984, Oper. Res..
[8] John N. Tsitsiklis,et al. The Complexity of Markov Decision Processes , 1987, Math. Oper. Res..
[9] Chelsea C. White,et al. Solution Procedures for Partially Observed Markov Decision Processes , 1989, Oper. Res..
[10] Hsien-Te Cheng,et al. Algorithms for partially observable markov decision processes , 1989 .
[11] William S. Lovejoy,et al. Computationally Feasible Bounds for Partially Observed Markov Decision Processes , 1991, Oper. Res..
[12] William S. Lovejoy,et al. Suboptimal Policies, with Bounds, for Parameter Adaptive Decision Processes , 1993, Oper. Res..
[13] Martin L. Puterman,et al. Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .
[14] Stuart J. Russell,et al. Approximating Optimal Policies for Partially Observable Stochastic Domains , 1995, IJCAI.
[15] Leslie Pack Kaelbling,et al. Learning Policies for Partially Observable Environments: Scaling Up , 1997, ICML.
[16] M. Littman,et al. Efficient dynamic-programming updates in partially observable Markov decision processes , 1995 .
[17] Michael L. Littman,et al. Algorithms for Sequential Decision Making , 1996 .
[18] Wenju Liu,et al. A Model Approximation Scheme for Planning in Partially Observable Stochastic Domains , 1997, J. Artif. Intell. Res..
[19] Milos Hauskrecht,et al. Planning and control in stochastic domains with imperfect information , 1997 .
[20] Michael L. Littman,et al. Incremental Pruning: A Simple, Fast, Exact Method for Partially Observable Markov Decision Processes , 1997, UAI.
[21] Milos Hauskrecht,et al. Incremental Methods for Computing Bounds in Partially Observable Markov Decision Processes , 1997, AAAI/IAAI.
[22] Ronen I. Brafman,et al. A Heuristic Variable Grid Solution Method for POMDPs , 1997, AAAI/IAAI.
[23] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..
[24] Eric A. Hansen,et al. Solving POMDPs by Searching in Policy Space , 1998, UAI.
[25] A. Cassandra,et al. Exact and approximate algorithms for partially observable markov decision processes , 1998 .
[26] Weihong Zhang,et al. A Method for Speeding Up Value Iteration in Partially Observable Markov Decision Processes , 1999, UAI.
[27] Milos Hauskrecht,et al. Value-Function Approximations for Partially Observable Markov Decision Processes , 2000, J. Artif. Intell. Res..
[28] Thomas G. Dietterich,et al. A POMDP Approximation Algorithm That Anticipates the Need to Observe , 2000, PRICAI.
[29] A. Cassandra. A Survey of POMDP Applications , 2003 .
[30] Eric V. Denardo,et al. Dynamic Programming: Models and Applications , 2003 .