Planning of actions presents a major computational challenge for animals as well as robots. Biological systems as well as AI systems must operate under high uncertainty. This uncertainty has many sources, for instance because of the unknown behavior of other animals or because of the noise induced by the limitations in the sensors. Thus, control of such systems is very different from an industrial robotic setting, where full knowledge can be assumed and noise can be almost ignored.

Optimizing a sequence of actions to attain some future goal is the general topic of control theory. The general stochastic control problem is intractable to solve and requires an exponential amount of memory and computation time. The reason is that the state space needs to be discretized and thus becomes exponentially large in the number of dimensions. Starting in 2004, we have proposed a novel class of stochastic control problems using path integrals that can be mapped onto a Bayesian inference problem. Because of its typical statistical mechanics form, one can consider various ways to approximate this path integral, such as the Laplace approximation, Monte Carlo sampling, mean field approximations or belief propagation.

Several robotics research groups world-wide have used this approach and shown that it significantly outperforms other state-of-the-art reinforcement methods. We are currently applying these methods coordinate the movement of Quadrotors (with UCL) and collaborate with Satoshi Satoh (ATR Japan), Takamitsu Matsubara (NAIST Japan) and Jan Peters (TU Darmstadt) on the specific issues related to the applicability of control theory to robotics.

The path integral theory makes quantitative predictions about optimal planning under uncertainty. One such prediction is the phenomenon of delayed choice: when uncertain about the future, it is wise, and optimal according to the theory, to delay a decision. The phenomenon is demonstrated in the Kappenball app (available on itunes.apple.com)

Current research aims at improving path integral control methods. We recently showed that the optimal control provides also an optimal sampling procedure. This provides a theoretical basis for the design of improved adaptive sampling schemes that 'learn' the optimal control. Furthermore, we provide novel theory that shows how to construct feed-back controllers with arbitrary complex state dependence.

IEEE transactions on Automatic Control,
vol. 62,
no. 3,
pp. 262-276.,
2017

Action selection in growing state spaces:
control of network structure growth.

Journal of Physics A,
vol. 50,
pp. 1-21,
2017
Particle smoothing for hidden diffusion processes: adaptive path integral smoother.

IEEE Transactions on Signal Processing,
vol. 65,
pp. 3391-3203,
2017
Efficient optical flow and stereo vision for an autonomous pocket drone..

IEEE transactions on Automatic Control,
vol. 2,
pp. 1070-1076,
2017
Journal of Statistical Mechanics: Theory and Experiment,
vol. 013206,
2016

All SNN publications