TU Darmstadt / ULB / TUprints

Nearest-Neighbor-based Collision Avoidance for Quadrotors via Reinforcement Learning

Ourari, Ramzi ; Cui, Kai ; Elshamanhory, Ahmed ; Koeppl, Heinz (2024)
Nearest-Neighbor-based Collision Avoidance for Quadrotors via Reinforcement Learning.
2022 IEEE International Conference on Robotics and Automation (ICRA). Philadelphia, PA, USA (23.05.2022-27.05.2022)
doi: 10.26083/tuprints-00028926
Conference or Workshop Item, Secondary publication, Postprint

[img] Text
Cui_et_al_2022_Nearest-Neighbor-based_Collision_Avoidance_for_Quadrotors_via_Reinforcement_Learning.pdf
Copyright Information: In Copyright.

Download (4MB)
Item Type: Conference or Workshop Item
Type of entry: Secondary publication
Title: Nearest-Neighbor-based Collision Avoidance for Quadrotors via Reinforcement Learning
Language: English
Date: 17 December 2024
Place of Publication: Darmstadt
Year of primary publication: 2022
Publisher: IEEE
Book Title: 2022 International Conference on Robotics and Automation (ICRA)
Event Title: 2022 IEEE International Conference on Robotics and Automation (ICRA)
Event Location: Philadelphia, PA, USA
Event Dates: 23.05.2022-27.05.2022
DOI: 10.26083/tuprints-00028926
Corresponding Links:
Origin: Secondary publication service
Abstract:

Collision avoidance algorithms are of central interest to many drone applications. In particular, decentralized approaches may be the key to enabling robust drone swarm solutions in cases where centralized communication becomes computationally prohibitive. In this work, we draw biological inspiration from flocks of starlings (Sturnus vulgaris) and apply the insight to end-to-end learned decentralized collision avoidance. More specifically, we propose a new, scalable observation model following a biomimetic nearest-neighbor information constraint that leads to fast learning and good collision avoidance behavior. By proposing a general reinforcement learning approach, we obtain an end-to-end learning-based approach to integrating collision avoidance with arbitrary tasks such as package collection and formation change. To validate the generality of this approach, we successfully apply our methodology through motion models of medium complexity, modeling momentum and nonetheless allowing direct application to real world quadrotors in conjunction with a standard PID controller. In contrast to prior works, we find that in our sufficiently rich motion model, nearest-neighbor information is indeed enough to learn effective collision avoidance behavior. Our learned policies are tested in simulation and subsequently transferred to real-world drones to validate their real-world applicability.

Uncontrolled Keywords: Automation, Computational modeling, Biological system modeling, Reinforcement learning, Behavioral sciences, Complexity theory, Collision avoidance
Status: Postprint
URN: urn:nbn:de:tuda-tuprints-289269
Classification DDC: 000 Generalities, computers, information > 004 Computer science
500 Science and mathematics > 570 Life sciences, biology
600 Technology, medicine, applied sciences > 621.3 Electrical engineering, electronics
Divisions: 18 Department of Electrical Engineering and Information Technology > Institute for Telecommunications > Bioinspired Communication Systems
18 Department of Electrical Engineering and Information Technology > Self-Organizing Systems Lab
Date Deposited: 17 Dec 2024 09:48
Last Modified: 17 Dec 2024 09:49
URI: https://tuprints.ulb.tu-darmstadt.de/id/eprint/28926
PPN:
Export:
Actions (login required)
View Item View Item