Jimma University Open access Institutional Repository

Machine Learning Based Energy-Efficient Uav- Assisted Emergency Wireless Communication For Information Collection In Post Disaster Area

Show simple item record

dc.contributor.author Mohammed Yesuf
dc.contributor.author Esmeal Kedir
dc.contributor.author Yekoye
dc.date.accessioned 2024-10-07T07:32:23Z
dc.date.available 2024-10-07T07:32:23Z
dc.date.issued 2023-10
dc.identifier.uri https://repository.ju.edu.et//handle/123456789/9286
dc.description.abstract Over the past few years, with the rapid increase in the number of natural disasters, the need to provide smart emergency wireless communication services to collect victim`s information has become crucial. Unmanned aerial Vehicles (UAVs) have gained much attention as promising candidates due to their unprecedented capabilities and broad flexibility. However, the limited UAVs battery and the limited UE`s energy, due to the breakdown of ground power system after disasters affected the mission, which necessitates an energy-efficient emergency wireless access procedure to insure the network lifetime. In this thesis, we propose multi-UAVs assisted emergency wireless access to collect disaster time information, where the UAVs fly safely to the victim`s UEs and control the UEs transmit power during the communication access. Oure goal is minimize the total energy consumption of both the UAVs and the UEs, which is needed to accomplish the emergency wireless access mission. We formulate this problem into three sub-problems of single UAV navigation, UE power control, and multi-UAV scheduling, and model each part as a finite-horizon Markov Decision Process (MDP). We deploy deep reinforcement learning (RL)-based frameworks to solve each part. Specifically, we use the deep deterministic policy gradient (DDPG) method to control the UE’s transmit power during wireless access and to generate the best trajectory for the UAVs in an obstacle-constrained environment to reach the UEs. To schedule activity plans for each UAVs to visit the UEs, we propose a multi-agent deep Q-learning (DQL) approach by taking the energy consumption of the UAVs on each path into account. Our simulations show that the UAVs can find a safe and optimal path for each of their trips. Continuous power control of the UE achieves better performance than the fixed power and fixed rate approaches in terms of the UE’s energy consumption and the wireless access completion time. In addition, compared to the two commonly used baselines framework, our multi-UAVs scheduling framework achieves better results in the simulated scenario. en_US
dc.language.iso en_US en_US
dc.subject unmanned aerial vehicle, trajectory optimization, deep reinforcement learning, multi-UAVs, Markov Decision Process, post-disaster en_US
dc.title Machine Learning Based Energy-Efficient Uav- Assisted Emergency Wireless Communication For Information Collection In Post Disaster Area en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search IR


Browse

My Account