Deep Reinforcement Learning for Drone Conflict Resolution Under Uncertainty
Deep Reinforcement Learning for Drone Conflict Resolution Under Uncertainty
| dc.contributor.author | Vlaskin, A.V. | |
| dc.contributor.author | Rahman, M.F. | |
| dc.contributor.author | Sunil, E. | |
| dc.contributor.author | Ellerbroek, J. | |
| dc.contributor.author | Hoekstra, J.M. | |
| dc.contributor.author | Nieuwenhuisen, D. | |
| dc.date.accessioned | 2026-01-12T14:20:56Z | |
| dc.date.available | 2026-01-12T14:20:56Z | |
| dc.date.issued | 2025 | |
| dc.description.abstract | Drone operations are expected to become a key facet of urban life, through contributions to emergency response missions, alleviation of urban traffic through parcel delivery, and aerial inspection. High demand is predicted to cause traffic densities never before seen in classical aviation, exceeding human capabilities and requiring autonomous solutions for separation management. In flight, drones are to perform maneuvers to avoid other vehicles - this is Conflict Resolution. In most research, the Conflict Detection and Resolution (CD&R) component is modeled without sensor error, which is unrealistic, as the State-Based conflict detection is to rely on GNSSderived position information. Previous work has shown that conventional methods such as the Modified Voltage Potential (MVP) are negatively impacted by this error. This paper examines the applicability and performance of Reinforcement Learning (RL) for this conflict detection and resolution task, with positional error modeled. The RL model outputs velocity and heading commands (actions) on the basis of its state information, and that of other vehicles. Two models are trained - one with, and one without gaussian noise applied to the observed vehicle position. Both of these are tested against the MVP algorithm as a benchmark. The difference between the RL model trained with and without noise is minor in terms of losses of separation (safety). Compared to MVP, the model consistently shows a lower loss of separation count, demonstrating better noise robustness. The RL models favor pure speed changes to resolve conflicts, while staying on the same course. | |
| dc.identifier.citation | A.V. Vlaskin, M.F. Rahman, E. Sunil, J. Ellerbroek, J.M. Hoekstra, D. Nieuwenhuisen, "Deep Reinforcement Learning for Drone Conflict Resolution Under Uncertainty", SESAR Innovation Days 2025, Bled, Slovenia | |
| dc.identifier.uri | https://hdl.handle.net/10921/1837 | |
| dc.language.iso | en | |
| dc.publisher | SESAR Joint Undertaking | |
| dc.title | Deep Reinforcement Learning for Drone Conflict Resolution Under Uncertainty | |
| dc.type | Other |
Files
Original bundle
1 - 1 of 1
- Name:
- SIDs_2025_paper_72-final.pdf
- Size:
- 3.67 MB
- Format:
- Adobe Portable Document Format
- Description: