Reinforcement learning: A survey LP Kaelbling, ML Littman, AW Moore Journal of artificial intelligence research 4, 237-285, 1996 | 8164 | 1996 |
Planning and acting in partially observable stochastic domains LP Kaelbling, ML Littman, AR Cassandra Artificial intelligence 101 (1-2), 99-134, 1998 | 4129 | 1998 |
Learning in embedded systems LP Kaelbling MIT press, 1993 | 863 | 1993 |
Acting optimally in partially observable stochastic domains AR Cassandra, LP Kaelbling, ML Littman Aaai 94, 1023-1028, 1994 | 824 | 1994 |
Learning policies for partially observable environments: Scaling up ML Littman, AR Cassandra, LP Kaelbling Machine Learning Proceedings 1995, 362-370, 1995 | 806 | 1995 |
Acting under uncertainty: Discrete Bayesian models for mobile-robot navigation AR Cassandra, LP Kaelbling, JA Kurien Proceedings of IEEE/RSJ International Conference on Intelligent Robots and …, 1996 | 696 | 1996 |
On the complexity of solving Markov decision problems ML Littman, TL Dean, LP Kaelbling arXiv preprint arXiv:1302.4971, 2013 | 611 | 2013 |
Hierarchical task and motion planning in the now LP Kaelbling, T Lozano-Pérez 2011 IEEE International Conference on Robotics and Automation, 1470-1477, 2011 | 499 | 2011 |
An architecture for intelligent reactive systems LP Kaelbling Reasoning about actions and plans, 395-410, 1987 | 486 | 1987 |
Effective reinforcement learning for mobile robots WD Smart, LP Kaelbling Proceedings 2002 IEEE International Conference on Robotics and Automation …, 2002 | 482 | 2002 |
The synthesis of digital machines with provable epistemic properties SJ Rosenschein, LP Kaelbling Theoretical aspects of reasoning about knowledge, 83-98, 1986 | 466 | 1986 |
To transfer or not to transfer MT Rosenstein, Z Marx, LP Kaelbling, TG Dietterich NIPS 2005 workshop on transfer learning 898, 1-4, 2005 | 385 | 2005 |
Input Generalization in Delayed Reinforcement Learning: An Algorithm and Performance Comparisons. D Chapman, LP Kaelbling IJCAI 91, 726-731, 1991 | 366 | 1991 |
Hierarchical solution of Markov decision processes using macro-actions M Hauskrecht, N Meuleau, LP Kaelbling, TL Dean, C Boutilier arXiv preprint arXiv:1301.7381, 2013 | 359 | 2013 |
Learning to cooperate via policy search L Peshkin, KE Kim, N Meuleau, LP Kaelbling arXiv preprint cs/0105032, 2001 | 346 | 2001 |
Action and planning in embedded agents LP Kaelbling, SJ Rosenschein Robotics and autonomous systems 6 (1-2), 35-48, 1990 | 340 | 1990 |
Planning under time constraints in stochastic domains T Dean, LP Kaelbling, J Kirman, A Nicholson Artificial Intelligence 76 (1-2), 35-74, 1995 | 329 | 1995 |
Practical reinforcement learning in continuous spaces WD Smart, LP Kaelbling ICML, 903-910, 2000 | 320 | 2000 |
Belief space planning assuming maximum likelihood observations R Platt Jr, R Tedrake, L Kaelbling, T Lozano-Perez | 305 | 2010 |
Learning topological maps with weak local odometric information H Shatkay, LP Kaelbling IJCAI (2), 920-929, 1997 | 293 | 1997 |