Title :
Basis iteration for reward based dimensionality reduction
Author_Institution :
Kalamazoo Coll., Kalamazoo
Abstract :
We propose a linear dimensionality reduction algorithm that selectively preserves task relevant state data for control problems modeled as Markov decision processes. The algorithm works by alternating value function estimation with basis vector adaptation. The approach is demonstrated on two tasks: a toy task designed to illustrate the key concepts, and a more complex three dimensional navigation task.
Keywords :
iterative methods; learning (artificial intelligence); Markov decision process; basis vector adaptation; iteration method; reward based dimensionality reduction; value function estimation; Data mining; Educational institutions; Independent component analysis; Learning; Least squares approximation; Least squares methods; Navigation; Predictive coding; Principal component analysis; Vectors; Dimensionality Reduction; Perceptual Development; Reinforcement Learning;
Conference_Titel :
Development and Learning, 2007. ICDL 2007. IEEE 6th International Conference on
Conference_Location :
London
Print_ISBN :
978-1-4244-1116-0
Electronic_ISBN :
978-1-4244-1116-0
DOI :
10.1109/DEVLRN.2007.4354032