References
- P. Wei et al., “Reinforcement Learning-Empowered Mobile Edge Computing for 6G Edge Intelligence,” IEEE Access, vol.10, pp. 65156-65192, 2022. doi: 10.1109/ACCESS.2022.3183647
- H. Sami, H. Otrok, J. Bentahar, and A. Mourad, “AI-based resource provisioning of IoE services in 6G: A deep reinforcement learning approach,” IEEE Trans. Network and Service Management, vol. 18, no. 3, pp. 3527-3540, 2021. doi:10.1109/TNSM.2021.3066625
- P. Zhou et al., “QoE aware 3D video streaming via deep reinforcement learning in software defined networking enabled mobile edge computing,” IEEE Trans. Netw. Sci. Eng., vol. 8, no. 1, pp. 419-433, 2021. doi:10.1109/TNSE.2020.3038998
- Y. Kunpeng et al., “Reinforcement learning-based mobile edge computing and transmission scheduling for video surveillance,” IEEE Trans. Emerg. Topics Comput, vol. 10, no. 2, pp. 1142-1156, 2021. doi: 10.1109/TETC.2021.3073744
- O. Yildiz and R. Sokullu, “Deep Q-Learning based resource allocation and load balancing in a mobile edge system serving different types of user requests,” Journal of Electrical Engineering, vol. 74, no. 1, pp. 48-55, 2023. doi: 10.2478/jee-2023-0005
- M. Rouissat et al., “Implementing and evaluating a new silent rank attack in RPL-contiki based IoT networks,” Journal of Electrical Engineering, vol.74, no. 6, pp. 454-462, 2023. doi: 10.2478/jee-2023-0053
- A. Khan, S. Gupta, and S. K. Gupta, “Multi-UAV integrated HetNet for maximum coverage in disaster management,” Journal of Electrical Engineering, vol. 73, no. 2, pp. 116-123, 2022. doi:10.2478/jee-2022-0015
- Q. Wu, Y. Zeng, and R. Zhang, “Joint trajectory and communication design for multi-UAV enabled wireless networks,” IEEE Transactions on Wireless Communications, vol. 17, no. 3, pp. 2109-2121, 2018. doi: 10.1109/TWC.2017.278929
- S. Amalorpava Mary Rajee and A. Merline, “Machine intelligence technique for blockage effects in next-generation heterogeneous networks,” Radioengineering, vol. 29, no. 3, 2020. doi: 10.13164/re.2020.0555
- S. Nagarajan et al., “Multi-Agent Reinforcement Learning for resource allocation in container based cloud environment,” Expert Syst., vol. 1, no. 22, 2023. https://doi.org/10.1111/exsy.13362
- V. T. M. Babu et al., “Survey on data communication for UAV and flight control system in MM wave communications,” In AIP Conference Proceedings, vol. 2790, no. 1, Aug. 2023. doi: 10.1063/5.0152781
- H. Sun, X. Chen, Q. Shi, M. Hong, X. Fu, and N. D. Sidiropoulos, “Learning to optimize: Training deep neural networks for interference management,” IEEE Trans. Signal Processing, vol. 66, no. 20, pp. 5438-5453, Oct 2018. doi:10.1109/TSP.2018.2866382
- M. Navaneethakrishnan et al., “Design of Biped Robot Using Reinforcement Learning and Asynchronous Actor-Critical Agent (A3C) Algorithm,” In 2023 2nd IEEE International Conference on Vision Towards Emerging Trends in Communication and Networking Technologies (ViTECoN), pp. 1-6, 2023.
- P. V. Klaine, M. A. Imran, O. Onireti, and R. D. Souza, “A survey of machine learning techniques applied to self-organizing cellular networks,” IEEE Commun. Surv. Tutor., vol. 19, no. 4, pp. 2392-2431, 2017. doi:10.1109/COMST.2017.2727878.
- R. Li et al., “Intelligent 5G: When cellular networks meet artificial intelligence,” IEEE Wirel. Commun.,” vol. 24, no. 5, pp. 175-183, 2017. doi: 10.1109/MWC.2017.1600304WC
- K. Guo, R. Gao, W. Xia, and T. Q. S. Quek, “Online learning based computation offloading in MEC systems with communication and computation dynamics,” IEEE Trans. Commun., vol. 69, no. 2, pp. 1147-1162, 2021. doi: 10.1109/TCOMM.2020.3038875
- S. A. M. Rajee, A. Merline, and M. M. Yamuna Devi, “Game theoretic model for power optimization in next- generation heterogeneous network,” SIViP, vol. 17, pp. 3721-3729, 2023. doi:10.1007/s11760-023-02599-8
- F. Ahmed, and M. Jenihhin, “A Survey on UAV Computing Platforms: A Hardware Reliability Perspective,” Sensors, vol. 22, no. 16, p. 6286. 2022. doi:10.3390/s22166286