Project Title Exploring the exploration problem Project Goal The control of a manipulator (such as a robot arm) in a non-deterministic environment has to take factors such as the device's dynamics, inertia, and time lags in the control signal flow into account. To learn the sensorimotor laws that will move the device on a desired trajectory, data has to be collected from exploratory movements in the environment. These exploratory movements need to be generated without yet knowing the precise sensorimotor laws and should reasonably cover the state space of the device. This project addresses the problem of how these exploratory movements can be generated. Going beyond simple exploration mechanisms such as random drifting or oscillating motor signals, it is the goal of this project to find theoretically grounded techniques that explore the state space in a way that is optimal from the learning point of view, i.e. the data collected in this ways leads to most efficient learning. Firstly, approaches to closely related problems in other areas will be explored. Among these are techniques for optimal data selection in statistical learning (active learning), model-based reinforcement learning approaches with standard behaviour planning techniques (dynamic programming), as well as guided and imitation learning for motor control. Secondly, the question could also be addressed analytically in a simple dynamic environment, e.g. a 2D variable with linear dynamics, assuming noisy data. References:
Project Timeline
This preliminary timeline will be updated with more detail as the research report and proposal are completed. Project Results & Conclusions |