Inhalt des Dokuments
robotics, the sensory inputs are high dimensional, however, only a
space is important to guide the actions of a robot. Previous work by Jonschkow-
ski and Brock in 2015 presented an unsupervised learning method extracting
a low dimensional representation using robotic priors. These robotic priors en-
code knowledge about the physical world in a loss function. However, their
and that of other methods to learn such a representation rely on an Marko-
vian observation space.
Description of Work
We extend the idea of robotic priors to work on
Markovian observation spaces. For this, we train a recurrent neural network
on trajectories, such that the network learns to encode past information in its
hidden state. With this we are able to learn a Markovian state space. To train
this network, we combine and modify existing robotic priors to work in non
Markovian environments. We test our method in a 3D maze environment. To
evaluate the quality of the learned state representation, we introduce a valida-
tion network that maps from the learned states to the ground truth.
new validation network, we show that the learned state space contains
positional and structural information of the environment. Furthermore, we use
reinforcement learning and show that the learned state space is sufficient to
solve a maze navigation task.