Learning-based Decentralized Control of Groups of Multiple Robots

Publié le
Equipe
Date de début de thèse (si connue)
1/9/2022
Lieu
Rennes
Unité de recherche
IRISA - UMR 6074
Description du sujet de la thèse

Over the past years many advancements have been made in devising decentralized control and estimation schemes for groups of multiple robots. From the initial consensus-like laws for solving simple rendez-vous or formation control problems [1], we have witnessed a flourishing or more complex algorithms able to embed and enforce much more complex "behaviors" for the group, such as inter-agent and obstacle collision avoidance, possibility of changing the sensing/communication topology at runtime, maintenance of "global" group properties such as connectivity or rigidity, implementation of navigation or exploration strategies in unknown environments, and interfacement with a human operator in charge of giving high-level commands to the group (or part of the group itself) [2,3,4,5]. While several theoretical frameworks have been proposed for embedding an arbitrary number of "complex" behaviors  in a single (but "arbitrarily complex") decentalized control law, this complexification has some costs, one of which being the increasing computational and communication complexity of the control law as a function of the "complexity" of the sought behaviors and of the group size.

One possible (and promising) alternative to the implementation of analytical (but computationally complex) decentalized laws is the use of modern learning techniques in the following illustrative pipeline: first, an analytical (or model-based) decentralized is designed for implementing the desired group behaviors. Then, a suitable neural network is trained on the model-based law in order to reproduce its effects. If the training is successful, the neural network can then be used in place of the model-based law with a key advantage: the computation complexity of the network is in general very law and scalable, contrarily to the model-based law, and thus much more amenable to be used on computationally-constrained robots such as small UAVs with limited onboard hardware [6,7,8].

We have recently started investigating these ideas with promising results in reproducing the behavior of a previously proposed decentralized connectivity maintenance law with inter-agent and obstacle avoidance requirements. The trained network could successfully reproduce the model-based controller in many simulated tests. The goal of this Thesis is start from this initial work and expand it in several directions, for instance: how to train the network in a decentralized way, how to include sensory data (images from onboard camera) in the learning, how to include the presence of humans either in the team or commanding the team. The activities will be tested experimentally on the quadrotor UAVs available in the Rainbow team.

Bibliographie

[1] M. Mesbahi and M. Egerstedt, Graph theoretic methods in multiagent networks. Princeton University Press, 2010.

[2] P. Robuffo Giordano, A. Franchi, C. Secchi, H. H. Bülthoff. A Passivity-Based Decentralized Strategy for Generalized Connectivity Maintenance. The International Journal of Robotics Research, 32(3):299-323, March 2013

[3] D. Zelazo, A. Franchi, H.-H. Bülthoff, P. Robuffo Giordano. Decentralized Rigidity Maintenance Control with Range Measurements for Multi-Robot Systems. The International Journal of Robotics Research, IJRR, 34(1):105-128, January 2015

[4] M. Aggravi, A. Alaaeldin Said Elsherif, P. Robuffo Giordano, C. Pacchierotti. Haptic-Enabled Decentralized Control of a Heterogeneous Human-Robot Team for Search and Rescue in Partially-known Environments. IEEE Robotics and Automation Letters (also presented at ICRA'21), 6(3):4843-4850, July 2021

[5] M. Aggravi, C. Pacchierotti, P. Robuffo Giordano. Connectivity-Maintenance Teleoperation of a UAV Fleet with Wearable Haptic Feedback. IEEE Trans. on Automation Science and Engineering, 18(3):1243-1262, June 2021

[6] V. Derhami and Y. Momeni, “Applying Reinforcement Learning in Formation Control of Agents,” in Intelligent Distributed Computing IX. Springer, 2016, pp. 297–307

[7] R. John and O. Andersson, “Intelligent formation control using reinforcement learning,” vol. Linkoping University, pp. 1–62, 01 2017.

[8] A. Geramifard, J. Redding, and J. P. How, “Intelligent cooperative control architecture: A framework for performance improvement using safe learning,” Journal of Intelligent & Robotic Systems, vol. 72, no. 1, pp. 83–103, Oct 2013

Liste des encadrants et encadrantes de thèse

Nom, Prénom
ROBUFFO GIORDANO, Paolo
Type d'encadrement
Directeur.trice de thèse
Unité de recherche
UMR 6074
Equipe

Nom, Prénom
PACCHIEROTTI, Claudio
Type d'encadrement
2e co-directeur.trice (facultatif)
Unité de recherche
UMR 6074
Equipe
Contact·s
Nom
ROBUFFO GIORDANO, Paolo
Email
prg@irisa.fr
Mots-clés
Learning, Multi-Robots, Decentralization, Human-Robot Interaction