Autonomous Navigation for Cellular-Connected UAV in Highly Dynamic Environments: A Deep Reinforcement Learning ApproachSource: Journal of Aerospace Engineering:;2024:;Volume ( 037 ):;issue: 005::page 04024063-1DOI: 10.1061/JAEEEZ.ASENG-5265Publisher: American Society of Civil Engineers
Abstract: This study investigated the navigation problem for cellular-connected unmanned aerial vehicles (UAVs), particularly in highly dynamic urban environments. To address this problem, the UAV is required not only to evade high-speed obstacles in the airspace but also to avoid the coverage holes of cellular base stations (BS). Moreover, the UAV needs to reach the destination to complete the navigation task. Hence, it is imperative to design the trade-off in action selections between collision evasion and destination-approaching scenarios, while also considering the expected communication outage duration as a crucial reference. To overcome this multiobjective optimization challenge, we propose a deep reinforcement learning (DRL)-based algorithm aimed at enabling the UAV to acquire an optimal decision-making policy. Specifically, we formulated the navigation problem as a Markov decision process (MDP) and developed a layered recurrent soft actor–critic (RSAC)-based DRL framework, stimulating the UAV to resolve two fundamental subtasks of UAV navigation. Furthermore, we develop a multilayer perception (MLP)-based integrated evaluation network to select a particular action from the two subsolutions, satisfying the demands for the entire navigation problem. The layered architecture simplifies the navigation problem, thereby enhancing the convergence speed of the proposed algorithm. Numerical results indicate that the layered-RSAC-based UAV can autonomously perform scheduled navigation tasks in our designed simulated urban environments with superior effectiveness.
|
Collections
Show full item record
contributor author | Di Wu | |
contributor author | Zhiyi Shi | |
contributor author | Yibo Zhang | |
contributor author | Mengxing Huang | |
date accessioned | 2024-12-24T10:14:06Z | |
date available | 2024-12-24T10:14:06Z | |
date copyright | 9/1/2024 12:00:00 AM | |
date issued | 2024 | |
identifier other | JAEEEZ.ASENG-5265.pdf | |
identifier uri | http://yetl.yabesh.ir/yetl1/handle/yetl/4298542 | |
description abstract | This study investigated the navigation problem for cellular-connected unmanned aerial vehicles (UAVs), particularly in highly dynamic urban environments. To address this problem, the UAV is required not only to evade high-speed obstacles in the airspace but also to avoid the coverage holes of cellular base stations (BS). Moreover, the UAV needs to reach the destination to complete the navigation task. Hence, it is imperative to design the trade-off in action selections between collision evasion and destination-approaching scenarios, while also considering the expected communication outage duration as a crucial reference. To overcome this multiobjective optimization challenge, we propose a deep reinforcement learning (DRL)-based algorithm aimed at enabling the UAV to acquire an optimal decision-making policy. Specifically, we formulated the navigation problem as a Markov decision process (MDP) and developed a layered recurrent soft actor–critic (RSAC)-based DRL framework, stimulating the UAV to resolve two fundamental subtasks of UAV navigation. Furthermore, we develop a multilayer perception (MLP)-based integrated evaluation network to select a particular action from the two subsolutions, satisfying the demands for the entire navigation problem. The layered architecture simplifies the navigation problem, thereby enhancing the convergence speed of the proposed algorithm. Numerical results indicate that the layered-RSAC-based UAV can autonomously perform scheduled navigation tasks in our designed simulated urban environments with superior effectiveness. | |
publisher | American Society of Civil Engineers | |
title | Autonomous Navigation for Cellular-Connected UAV in Highly Dynamic Environments: A Deep Reinforcement Learning Approach | |
type | Journal Article | |
journal volume | 37 | |
journal issue | 5 | |
journal title | Journal of Aerospace Engineering | |
identifier doi | 10.1061/JAEEEZ.ASENG-5265 | |
journal fristpage | 04024063-1 | |
journal lastpage | 04024063-14 | |
page | 14 | |
tree | Journal of Aerospace Engineering:;2024:;Volume ( 037 ):;issue: 005 | |
contenttype | Fulltext |