Smart City Gnosys

Smart city article details

Title A Lightweight Reinforcement-Learning-Based Real-Time Path-Planning Method For Unmanned Aerial Vehicles
ID_Doc 2337
Authors Xi M.; Dai H.; He J.; Li W.; Wen J.; Xiao S.; Yang J.
Year 2024
Published IEEE Internet of Things Journal, 11, 12
DOI http://dx.doi.org/10.1109/JIOT.2024.3350525
Abstract The unmanned aerial vehicles (UAVs) are competent to perform a variety of applications, possessing great potential and promise. The deep neural networks (DNN) technology has enabled the UAV-assisted paradigm, accelerated the construction of smart cities, and propelled the development of the Internet of Things (IoT). UAVs play an increasingly important role in various applications, such as surveillance, environmental monitoring, emergency rescue, and supplies delivery, for which a robust path-planning technique is the foundation and prerequisite. However, existing methods lack comprehensive consideration of the complicated urban environment and do not provide an overall assessment of the robustness and generalization. Meanwhile, due to the resource constraints and hardware limitations of UAVs, the complexity of deploying the network needs to be reduced. This article proposes a lightweight, reinforcement-learning-based (RL-based) real-time path-planning method for UAVs, adaptive SAC (ASAC) algorithm, which optimizing the training process, network architecture, and algorithmic models. First of all, we establish a framework of global training and local adaptation, where the structured environment model is constructed for interaction, and local dynamically varying information aids in improving generalization. Second, ASAC introduces a cross-layer connection approach that passes the original state information into the higher layers to avoid feature loss and improve learning efficiency. Finally, we propose an adaptive temperature coefficient, which flexibly adjusts the exploration probability of UAVs with the training phase and experience data accumulation. In addition, a series of comparison experiments have been conducted in conjunction with practical application requirements, and the results have fully proved the favorable superiority of ASAC. © 2014 IEEE.
Author Keywords Internet of Things (IoT); path-planning; realtime applications; reinforcement-learning; tiny machine learning (ML); unmanned aerial vehicles (UAVs)


Similar Articles


Id Similarity Authors Title Published
9695 View0.877Chen S.; Wei K.; Pei T.; Long S.Aoi-Guaranteed Uav Crowdsensing: A Ugv-Assisted Deep Reinforcement Learning ApproachAd Hoc Networks, 173 (2025)
6398 View0.874Qadir, Z; Ullah, F; Munawar, HS; Al-Turjman, FAddressing Disasters In Smart Cities Through Uavs Path Planning And 5G Communications: A Systematic ReviewCOMPUTER COMMUNICATIONS, 168 (2021)
38095 View0.874Dhuheir M.; Erbad A.; Hamdaoui B.; Belhaouari S.B.; Guizani M.; Vu T.X.Multi-Agent Meta Reinforcement Learning For Reliable And Low-Latency Distributed Inference In Resource-Constrained Uav SwarmsIEEE Access, 13 (2025)
59297 View0.872Deng C.; Fang X.; Wang X.Uav-Enabled Mobile-Edge Computing For Ai Applications: Joint Model Decision, Resource Allocation, And Trajectory OptimizationIEEE Internet of Things Journal, 10, 7 (2023)
34952 View0.865Gao T.; Goins D.; Ballotti C.; Liu J.; Qu C.Learning-Based Uav Swarm Video Analytics Orchestration In Disaster Response ManagementSN Computer Science, 6, 5 (2025)
8277 View0.864Li C.; Feng Q.; Ding C.; Ye Z.An Improved Multi-Actor Hybrid Attention Critic Algorithm For Cooperative Navigation In Urban Low-Altitude Logistics EnvironmentsComputers, Materials and Continua, 84, 2 (2025)
18022 View0.861Chen L.; Liu K.; Yang P.; Xiong Z.; An P.; Quek T.Q.S.; Zhang Z.Deep Point Reinforcement Learning Approach For Sustainable Communications By Uav And Moving Interaction StationIEEE Transactions on Vehicular Technology (2025)
58717 View0.859Tao X.; Hafid A.S.Trajectory Design In Uav-Aided Mobile Crowdsensing: A Deep Reinforcement Learning ApproachIEEE International Conference on Communications (2021)
5700 View0.858Oubbati O.S.; Alotaibi J.; Alromithy F.; Atiquzzaman M.; Altimania M.R.A Uav-Ugv Cooperative System: Patrolling And Energy Management For Urban MonitoringIEEE Transactions on Vehicular Technology (2025)
2202 View0.857Ulaş M.; Sezgin A.; Boyacı A.A Hybrid Optimization Framework For Dynamic Drone Networks: Integrating Genetic Algorithms With Reinforcement LearningApplied Sciences (Switzerland), 15, 9 (2025)