Ourari, Ramzi ; Cui, Kai ; Elshamanhory, Ahmed ; Koeppl, Heinz (2024)
Nearest-Neighbor-based Collision Avoidance for Quadrotors via Reinforcement Learning.
2022 IEEE International Conference on Robotics and Automation (ICRA). Philadelphia, PA, USA (23.05.2022-27.05.2022)
doi: 10.26083/tuprints-00028926
Conference or Workshop Item, Secondary publication, Postprint
Text
Cui_et_al_2022_Nearest-Neighbor-based_Collision_Avoidance_for_Quadrotors_via_Reinforcement_Learning.pdf Copyright Information: In Copyright. Download (4MB) |
Item Type: | Conference or Workshop Item |
---|---|
Type of entry: | Secondary publication |
Title: | Nearest-Neighbor-based Collision Avoidance for Quadrotors via Reinforcement Learning |
Language: | English |
Date: | 17 December 2024 |
Place of Publication: | Darmstadt |
Year of primary publication: | 2022 |
Publisher: | IEEE |
Book Title: | 2022 International Conference on Robotics and Automation (ICRA) |
Event Title: | 2022 IEEE International Conference on Robotics and Automation (ICRA) |
Event Location: | Philadelphia, PA, USA |
Event Dates: | 23.05.2022-27.05.2022 |
DOI: | 10.26083/tuprints-00028926 |
Corresponding Links: | |
Origin: | Secondary publication service |
Abstract: | Collision avoidance algorithms are of central interest to many drone applications. In particular, decentralized approaches may be the key to enabling robust drone swarm solutions in cases where centralized communication becomes computationally prohibitive. In this work, we draw biological inspiration from flocks of starlings (Sturnus vulgaris) and apply the insight to end-to-end learned decentralized collision avoidance. More specifically, we propose a new, scalable observation model following a biomimetic nearest-neighbor information constraint that leads to fast learning and good collision avoidance behavior. By proposing a general reinforcement learning approach, we obtain an end-to-end learning-based approach to integrating collision avoidance with arbitrary tasks such as package collection and formation change. To validate the generality of this approach, we successfully apply our methodology through motion models of medium complexity, modeling momentum and nonetheless allowing direct application to real world quadrotors in conjunction with a standard PID controller. In contrast to prior works, we find that in our sufficiently rich motion model, nearest-neighbor information is indeed enough to learn effective collision avoidance behavior. Our learned policies are tested in simulation and subsequently transferred to real-world drones to validate their real-world applicability. |
Uncontrolled Keywords: | Automation, Computational modeling, Biological system modeling, Reinforcement learning, Behavioral sciences, Complexity theory, Collision avoidance |
Status: | Postprint |
URN: | urn:nbn:de:tuda-tuprints-289269 |
Classification DDC: | 000 Generalities, computers, information > 004 Computer science 500 Science and mathematics > 570 Life sciences, biology 600 Technology, medicine, applied sciences > 621.3 Electrical engineering, electronics |
Divisions: | 18 Department of Electrical Engineering and Information Technology > Institute for Telecommunications > Bioinspired Communication Systems 18 Department of Electrical Engineering and Information Technology > Self-Organizing Systems Lab |
Date Deposited: | 17 Dec 2024 09:48 |
Last Modified: | 17 Dec 2024 09:49 |
URI: | https://tuprints.ulb.tu-darmstadt.de/id/eprint/28926 |
PPN: | |
Export: |
View Item |