DocumentCode :
2625261
Title :
A formal framework for robot learning and control under model uncertainty
Author :
Jaulmes, Robin ; Pineau, Joelle ; Precup, Doina
Author_Institution :
Sch. of Comput. Sci., McGill Univ., Montreal, Que.
fYear :
2007
fDate :
10-14 April 2007
Firstpage :
2104
Lastpage :
2110
Abstract :
While the partially observable Markov decision process (POMDP) provides a formal framework for the problem of robot control under uncertainty, it typically assumes a known and stationary model of the environment. In this paper, we study the problem of finding an optimal policy for controlling a robot in a partially observable domain, where the model is not perfectly known, and may change over time. We present an algorithm called MEDUSA which incrementally learns a POMDP model using queries, while still optimizing a reward function. We demonstrate effectiveness of the approach for a simple scenario, where a robot seeking a person has minimal a priori knowledge of its own sensor model, as well as where the person is located.
Keywords :
Markov processes; robots; MEDUSA algorithm; model uncertainty; partially observable Markov decision process; robot learning; Automatic control; Computer science; Mobile robots; Optimal control; Robot control; Robot sensing systems; Robotics and automation; Sensor phenomena and characterization; Uncertainty; Wheelchairs;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Robotics and Automation, 2007 IEEE International Conference on
Conference_Location :
Roma
ISSN :
1050-4729
Print_ISBN :
1-4244-0601-3
Electronic_ISBN :
1050-4729
Type :
conf
DOI :
10.1109/ROBOT.2007.363632
Filename :
4209396
Link To Document :
بازگشت