Resource Optimization of Airborne Base Stations Using Artificial Intelligence Methods
- Authors: Tran T.D.1, Koucheryavy A.E.1
-
Affiliations:
- The Bonch-Bruevich Saint Petersburg State University of Telecommunications
- Issue: Vol 11, No 1 (2025)
- Pages: 62-68
- Section: ELECTRONICS, PHOTONICS, INSTRUMENTATION AND COMMUNICATIONS
- URL: https://journals.rcsi.science/1813-324X/article/view/283900
- DOI: https://doi.org/10.31854/1813-324X-2025-11-1-62-68
- EDN: https://elibrary.ru/RVENVC
- ID: 283900
Cite item
Full Text
Abstract
In remote areas and disaster-stricken regions, unmanned aerial vehicles (UAVs) can serve as base stations, providing wireless communication to ground users. Due to their high mobility, low cost, and rapid deployment and retrieval capabilities, UAVs can continuously adjust their position in three-dimensional (3D) space, improving wireless connectivity and enhancing data transmission rates. In this paper, we investigate the problem of ABS (Aerial Base Station) deployment in 3D space and power allocation with the aim of maximizing the data transmission rate in the system. To address this non-convex problem, we propose Q-learning, a reinforcement learning algorithm. By using the ABS as an agent, the algorithm enables the ABS to explore the state space and take actions based on an ϵ-greedy policy (optimal epsilon value) to determine its 3D position and power allocation. Simulation results demonstrate that the proposed algorithm outperforms individual position optimization and power allocation optimization.The purpose of this article is to study the efficiency of using modern artificial intelligence methods to optimize the use of resources of airborne base stations of public communication networks.The essence of the proposed solution is to use modern artificial intelligence methods, namely: the Q-learning method and the epsilon-greedy ϵ-greedy algorithm to ensure joint optimization of the placement of airborne base stations and power distribution to maximize the data transfer rate. The system has an implementation in the form of a simulation program. Simulation experiments have shown that the use of the Q-learning reinforcement learning method and the epsilon-greedy e-greedy algorithm for joint optimization provides a higher overall data transfer rate in the system compared to optimizing only the location or power distribution.The scientific novelty of the proposed solution is that joint optimization of the placement of an airborne base station and power distribution made it possible, in contrast to known results, to establish that the flight altitude of a UAV with a base station installed on it when optimizing only the location will be higher than the flight altitude of a UAV when jointly optimizing the location and power distribution.The practical significance is the possibility of developing a methodology for planning public communication networks using airborne base stations to obtain a higher overall data transfer rate on the corresponding network fragment.
Keywords
About the authors
T. D. Tran
The Bonch-Bruevich Saint Petersburg State University of Telecommunications
Email: chan.tz@sut.ru
ORCID iD: 0009-0006-0080-9477
A. E. Koucheryavy
The Bonch-Bruevich Saint Petersburg State University of Telecommunications
Email: akouch@sut.ru
ORCID iD: 0000-0003-4479-2479
SPIN-code: 1012-4238
References
- Ding G., Wu Q., Zhang L., Lin Y., Tsiftsis T.A., Yao Y.D. An amateur drone surveillance system based on the cognitive Internet of Things // IEEE Communications Magazine. 2018. Vol. 56. Iss.1. PP. 29‒35. doi: 10.1109/MCOM.2017.1700452. EDN:YBEOMH
- Rose K., Eldridge S., Chapin L. The internet of things: An overview // The internet society (ISOC). 2015. Vol. 80(15). PP. 1‒53.
- Zhao N., Lu W., Sheng M., Chen Y., Tang J., Yu F.R., et al. UAV-Assisted Emergency Networks in Disasters // IEEE Wireless Communications. 2019. Vol. 26. Iss. 1. PP. 45‒51. doi: 10.1109/MWC.2018.1800160
- Li B., Fei Z., Zhang Y. UAV communications for 5G and beyond: Recent advances and future trends // IEEE Internet of Things Journal. 2018. Vol. 6. Iss. 2. PP. 2241‒2263. doi: 10.1109/JIOT.2018.2887086. EDN:UPZMGQ
- Shannon C.E. A mathematical theory of communication // The Bell System Technical Journal. 1948. Vol. 27. Iss. 3. PP. 379‒423. doi: 10.1002/j.1538-7305.1948.tb01338.x
- Sutton R.S. Reinforcement learning: An introduction. The MIT Press, 2018.
- Jaakkola T., Jordan M., Singh S. Convergence of Stochastic Iterative Dynamic Programming Algorithms // Advances in Neural Information Processing Systems 6 (NIPS 1993). 1993.
Supplementary files

