dc.contributor.author |
Ebrahimi, Dariush |
|
dc.contributor.author |
Sharafeddine, Sanaa |
|
dc.contributor.author |
Ho, Pin-Han |
|
dc.contributor.author |
Assi, Chadi |
|
dc.date.accessioned |
2020-02-04T13:45:45Z |
|
dc.date.available |
2020-02-04T13:45:45Z |
|
dc.date.copyright |
2020 |
en_US |
dc.date.issued |
2020-02-04 |
|
dc.identifier.issn |
1536-1233 |
en_US |
dc.identifier.uri |
http://hdl.handle.net/10725/11798 |
|
dc.description.abstract |
Disaster management, search and rescue missions, and health monitoring are examples of critical applications that require object localization with high precision and sometimes in a timely manner. In the absence of the global positioning system (GPS), the radio received signal strength index (RSSI) can be used for localization purposes due to its simplicity and cost-effectiveness. However, due to the low accuracy of RSSI, unmanned aerial vehicles (UAVs) or drones may be used as an efficient solution for improved localization accuracy due to their agility and higher probability of line-of-sight (LoS). Hence, in this context, we propose a novel framework based on reinforcement learning (RL) to enable a UAV (agent) to autonomously find its trajectory that results in improving the localization accuracy of multiple objects in shortest time and path length, fewer signal-strength measurements (waypoints), and/or lower UAV energy consumption. In particular, we first control the agent through initial scan trajectory on the whole region to 1) know the number of nodes and estimate their initial locations, and 2) train the agent online during operation. Then, the agent forms its trajectory by using RL to choose the next waypoints in order to minimize the average location errors of all objects. Our framework includes detailed UAV to ground channel characteristics with an empirical path loss and log-normal shadowing model, and also with an elaborate energy consumption model. We investigate and compare the localization precision of our approach with existing methods from the literature by varying the UAV's trajectory length, energy, number of waypoints, and time. Furthermore, we study the impact of the UAV's velocity, altitude, hovering time, communication range, number of maximum RSSI measurements, and number of objects. The results show the superiority of our method over the state-of-art and demonstrates its fast reduction of the localization error. |
en_US |
dc.language.iso |
en |
en_US |
dc.title |
Autonomous UAV Trajectory for Localizing Ground Objects |
en_US |
dc.type |
Article |
en_US |
dc.description.version |
Published |
en_US |
dc.title.subtitle |
A Reinforcement Learning Approach |
en_US |
dc.author.school |
SAS |
en_US |
dc.author.idnumber |
200502746 |
en_US |
dc.author.department |
Computer Science And Mathematics |
en_US |
dc.description.embargo |
N/A |
en_US |
dc.relation.journal |
IEEE Transactions on Mobile Computing |
en_US |
dc.journal.volume |
20 |
|
dc.journal.issue |
4 |
|
dc.article.pages |
1312-1324 |
|
dc.keywords |
Localization |
en_US |
dc.keywords |
Reinforcement Learning |
en_US |
dc.keywords |
Q-Learning |
en_US |
dc.keywords |
Unmanned Aerial Vehicles (UAVs) |
en_US |
dc.keywords |
Drones |
en_US |
dc.keywords |
Trajectory Planning |
en_US |
dc.keywords |
Received Signal Strength (RSS) |
en_US |
dc.identifier.doi |
https://doi.org/10.1109/TMC.2020.2966989 |
en_US |
dc.identifier.ctation |
Ebrahimi, D., Sharafeddine, S., Ho, P. H., & Assi, C. (2020). Autonomous UAV Trajectory for Localizing Ground Objects: A Reinforcement Learning Approach. IEEE Transactions on Mobile Computing, 20 (4), 1312-1324. |
en_US |
dc.author.email |
sanaa.sharafeddine@lau.edu.lb |
en_US |
dc.identifier.tou |
http://libraries.lau.edu.lb/research/laur/terms-of-use/articles.php |
en_US |
dc.identifier.url |
https://ieeexplore.ieee.org/abstract/document/8960453 |
en_US |
dc.orcid.id |
https://orcid.org/0000-0001-6548-1624 |
en_US |
dc.author.affiliation |
Lebanese American University |
en_US |