Deep Reinforcement Learning for Drone Conflict Resolution Under Uncertainty

dc.contributor.author Vlaskin, A.V.
dc.contributor.author Rahman, M.F.
dc.contributor.author Sunil, E.
dc.contributor.author Ellerbroek, J.
dc.contributor.author Hoekstra, J.M.
dc.contributor.author Nieuwenhuisen, D.
dc.date.accessioned 2026-01-12T14:20:56Z
dc.date.available 2026-01-12T14:20:56Z
dc.date.issued 2025
dc.description.abstract Drone operations are expected to become a key facet of urban life, through contributions to emergency response missions, alleviation of urban traffic through parcel delivery, and aerial inspection. High demand is predicted to cause traffic densities never before seen in classical aviation, exceeding human capabilities and requiring autonomous solutions for separation management. In flight, drones are to perform maneuvers to avoid other vehicles - this is Conflict Resolution. In most research, the Conflict Detection and Resolution (CD&R) component is modeled without sensor error, which is unrealistic, as the State-Based conflict detection is to rely on GNSSderived position information. Previous work has shown that conventional methods such as the Modified Voltage Potential (MVP) are negatively impacted by this error. This paper examines the applicability and performance of Reinforcement Learning (RL) for this conflict detection and resolution task, with positional error modeled. The RL model outputs velocity and heading commands (actions) on the basis of its state information, and that of other vehicles. Two models are trained - one with, and one without gaussian noise applied to the observed vehicle position. Both of these are tested against the MVP algorithm as a benchmark. The difference between the RL model trained with and without noise is minor in terms of losses of separation (safety). Compared to MVP, the model consistently shows a lower loss of separation count, demonstrating better noise robustness. The RL models favor pure speed changes to resolve conflicts, while staying on the same course.
dc.identifier.citation A.V. Vlaskin, M.F. Rahman, E. Sunil, J. Ellerbroek, J.M. Hoekstra, D. Nieuwenhuisen, "Deep Reinforcement Learning for Drone Conflict Resolution Under Uncertainty", SESAR Innovation Days 2025, Bled, Slovenia
dc.identifier.uri https://hdl.handle.net/10921/1837
dc.language.iso en
dc.publisher SESAR Joint Undertaking
dc.title Deep Reinforcement Learning for Drone Conflict Resolution Under Uncertainty
dc.type Other
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
SIDs_2025_paper_72-final.pdf
Size:
3.67 MB
Format:
Adobe Portable Document Format
Description: