DLR Portal
Home|Sitemap|Contact|Accessibility Imprint and terms of use Privacy Cookies & Tracking |Deutsch
You are here: Home:Organization:Communications Systems:Swarm Exploration:Swarm Exploration with Reinforcement Learning
Advanced Search
Our Missions
Highlights of Research
Organization
Satellite Networks
Optical Satellite Links
Communications Systems
Aeronautical Communications
Vehicular Applications
Multimodal Personal Navigation
Mobile Radio Transmission
Swarm Exploration
Navigation
Nautical Systems
Institute Project Management and Administration
Committees for participation
Staff
Laboratories and large-scale facilities
Projects
Job Offers
Scientific Publications
Multimedia
Back
Print

Swarm Exploration with Reinforcement Learning



Nowadays approaches to accomplish exploration tasks exploit a model that describes the environment and process of interest. This setup works fine as long as the model describes them precisely. In case the environment or the process changes, a new model has to be built and the algorithm must be adapted to it. This increases the effort required to develop algorithms for novel exploration tasks.

The Swarm Exploration group is developing machine learning algorithms that enable a swarm of robots to learn how to carry out complex exploration tasks. In particular, we focus here on model-free deep reinforcement learning (Deep-RL) approaches, which do not require a model of the environment and process of interest. RL algorithms are a family of algorithms that permit an agent to learn how to behave by interacting with the environment. This is done through a reward signal, which encodes how well the agent is performing. Hence, the aim of a RL agent is to learn a policy so as to maximize the expected future reward.


Model-free RL has been shown to offer outstanding results for a wide variety of tasks. Nevertheless, there are many applications for which a model of the process of interest has been well studied. This is the case, e.g., in one of our applications of interest: gas source localization. In gas source localization, partial differential equations have been proven to model the gas dispersion very accurately. Therefore, one of the questions that we also address in our research is: how can we introduce domain knowledge -- a model -- of a physical process in RL to solve an exploration task?


We developed a framework – DeepIG – that allows multiple robots to learn how to accomplish complex exploration tasks using Deep RL. In particular, our focus lies on terrain mapping, wildfire monitoring, and gas source localization tasks.

 

 

Videos

DeepIG: Multi-Robot Information Gathering with Deep Reinforcement Learning

Publications

Viseras, Alberto and Garcia, Ricardo (2019). Deepig: Multi-robot information gathering with deep reinforcement learning. IEEE Robotics and Automation Letters, 4(3), 3059-3066.

 


Contact
Juan Marchal Gomez
German Aerospace Center

Institute of Communications and Navigation
, Communications Systems
Oberpfaffenhofen-Wessling

Tel.: +49 8153 28-2676

Fax: +49 8153 28 2676

Related Topics
Cybernetics, Artificial Intelligence and Robotics
Numerical Analysis
Copyright © 2023 German Aerospace Center (DLR). All rights reserved.