Amir-massoud Farahmand
Amir-massoud Farahmand
Vector Institute
Verified email at vectorinstitute.ai - Homepage
TitleCited byYear
Regularized policy iteration
A Farahmand, M Ghavamzadeh, S Mannor, C Szepesvári
Advances in Neural Information Processing Systems, 441-448, 2009
1342009
Error propagation for approximate policy and value iteration
A Farahmand, C Szepesvári, R Munos
Advances in Neural Information Processing Systems, 568-576, 2010
842010
Manifold-adaptive dimension estimation
A Farahmand, C Szepesvári, JY Audibert
Proceedings of the 24th international conference on Machine learning, 265-272, 2007
74*2007
Regularized fitted Q-iteration for planning in continuous-space Markovian decision problems
A Farahmand, M Ghavamzadeh, C Szepesvári, S Mannor
2009 American Control Conference, 725-730, 2009
682009
Robust jacobian estimation for uncalibrated visual servoing
A Shademan, A Farahmand, M Jägersand
Robotics and Automation (ICRA), 2010 IEEE International Conference on, 5564-5569, 2010
502010
Global visual-motor estimation for uncalibrated visual servoing
A Farahmand, A Shademan, M Jagersand
2007 IEEE/RSJ International Conference on Intelligent Robots and Systems …, 2007
422007
Learning from Limited Demonstrations
B Kim, A Farahmand, J Pineau, D Precup
Advances in Neural Information Processing Systems (NIPS), 2859-2867, 2013
392013
Model Selection in Reinforcement Learning
AM Farahmand, C Szepesvári
Machine learning 85 (3), 299-332, 2011
302011
Model-based and model-free reinforcement learning for visual servoing
A massoud Farahmand, A Shademan, M Jagersand, C Szepesvári
2009 IEEE International Conference on Robotics and Automation, 2917-2924, 2009
202009
Regularized fitted Q-iteration: Application to planning
AM Farahmand, M Ghavamzadeh, C Szepesvári, S Mannor
Recent Advances in Reinforcement Learning, 55-68, 2008
20*2008
Regularized Fitted Q-iteration: Application to Bounded Resource Planning
A Farahmand, M Ghavamzadeh, C Szepesvári, S Mannor
20*
Approximate MaxEnt Inverse Optimal Control and its Application for Mental Simulation of Human Interactions
DA Huang, AM Farahmand, KM Kitani, JA Bagnell
AAAI Conference on Artificial Intelligence (AAAI), 2015
192015
Action-Gap Phenomenon in Reinforcement Learning
AM Farahmand
Neural Information Processing Systems (NIPS), 2011
172011
Value Pursuit Iteration
A Farahmand, D Precup
Advances in Neural Information Processing Systems 25, 1349-1357, 2012
162012
Regularization in Reinforcement Learning
AM Farahmand
University of Alberta, 2011
162011
Regularized policy iteration with nonparametric function spaces
A Farahmand, M Ghavamzadeh, C Szepesvári, S Mannor
The Journal of Machine Learning Research 17 (1), 4809-4874, 2016
152016
Bellman Error Based Feature Generation using Random Projections on Sparse Spaces
MM Fard, Y Grinberg, A Farahmand, J Pineau, D Precup
Advances in Neural Information Processing Systems (NIPS), 3030--3038, 2013
152013
Interaction of culture-based learning and cooperative co-evolution and its application to automatic behavior-based system design
AM Farahmand, MN Ahmadabadi, C Lucas, BN Araabi
Evolutionary Computation, IEEE Transactions on 14 (1), 23-57, 2010
152010
Towards learning robotic reaching and pointing: An uncalibrated visual servoing approach
A Shademan, A Farahmand, M Jagersand
2009 Canadian Conference on Computer and Robot Vision, 229-236, 2009
102009
Regularized least-squares regression: Learning from a [beta]-mixing sequence
AM Farahmand, C Szepesvári
Journal of Statistical Planning and Inference 142 (2), 493–505, 2012
92012
The system can't perform the operation now. Try again later.
Articles 1–20