[1] 孙 湧, 仵 博, 冯延蓬. 基于策略迭代和值迭代的POMDP算法[J]. 计算机研究与发展, 2008, 45(10): 1763-1768. [2] Kurniawati H, Hsu D, Lee W S. SARSOP: Efficient Point- based POMDP Planning by Approximating Optimally Reachable Belief Spaces[C]//Proc. of Robotics: Science and Systems. Zurich, Switzerland: MIT Press, 2008. [3] Ross S, Pineau J, Paquet S, et al. Online Planning Algorithms for POMDPs[J]. Journal of Artificial Intelligence Research, 2008, 32(1): 663-704. [4] 仵 博, 吴 敏, 佘锦华. 基于点的POMDPs在线值迭代算法[J]. 软件学报, 2013, 24(1): 25-36. [5] He Ruijie, Brunskill E, Roy N. Efficient Planning Under Uncertainty with Macro-actions[J]. Journal of Artificial Intelligence Research, 2011, 40(1): 523-570. [6] 仵 博, 吴 敏. 部分可观察马尔可夫决策过程研究进展[J]. 计算机工程与设计, 2007, 28(9): 2116-2119. [7] Boyen X, Koller D. Tractable Inference for Complex Stochastic Processes[C]//Proc. of the 14th Conference on Uncertainty in Artificial Intelligence. Madison, USA: Morgan Kaufmann Press, 1998. [8] Cohn R, Durfee E, Singh S. Planning Delayed-response Queries and Transient Policies Under Reward Uncertainty[C]//Proc. of the 7th Annual Workshop on Multiagent Sequential Decision-making Under Uncertainty. Valencia, Spain: ACM Press, 2012. [9] 周红芳, 李红岩, 刘 颖, 等. 多维数据集中聚类数确定算法研究[J]. 计算机工程, 2012, 38(9): 8-11. [10] Andrieu C, Doucet A, Holenstein R. Particle Markov Chain Monte Carlo Methods[J]. Journal of the Royal Statistical Society: Series B, 2010, 72(3): 269-342. [11] Kwok C, Fox D, Meila M. Real-time Particle Filters[J]. Proceedings of the IEEE, 2004, 92(3): 469-484. [12] Smith T, Simmons R. Point-based POMDP Algorithms: Improved Analysis and Implementation[C]//Proc. of the 21st Conference on Uncertainty in Artificial Intelligence. Arlington, USA: AUAI Press, 2005. 编辑 刘 冰
|