Abstract
Mobile edge computing (MEC) can dispatch its powerful servers close by to assist with the computation workloads that intelligent wireless terminals have offloaded. The MEC server’s physical location is closer to the intelligent wireless terminals, which can satisfy the low latency and high reliability demands. In this paper, we formulate an MEC framework with multiple vehicles and service devices that considers the priority and randomness of arriving workloads from roadside units (RSUs), cameras, laser radars (Lidar) and the time-varying channel state between the service device and MEC server (MEC-S). To minimize the long-term weighted average cost of the proposed MEC system, we transit this issue (cost minimization problem) into the Markov decision process (MDP). Furthermore, considering the difficulty realizing the state transition probability matrix, the dimensional complexity of the state space, and the continuity of the action space, we propose a deterministic policy gradient (MADDPG)-based bandwidth partition and power allocation optimization policy. The proposed MADDPG-based policy is a model-free deep reinforcement learning (DRL) method, which can effectively deal with continuous action space and further guide multi-agent to execute decision-making. The comprehensive results verify that the proposed MADDPG-based optimization scheme has fine convergence and performance that is better than that of the other four baseline algorithms.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Lakhan A, Ahmad M, Bilal M, Jolfaei A, Mehmood RM (2021) Mobility aware blockchain enabled offloading and scheduling in vehicular fog cloud computing. IEEE Trans Intell Transp Syst 22(7):4212–4223
Zhang X, Wang Y (2022) Deepmecagent: multi-agent computing resource allocation for uav-assisted mobile edge computing in distributed iot system. Appl Intell 1–12
Ding Y, Li K, Liu C, Li K (2021) A potential game theoretic approach to computation offloading strategy optimization in end-edge-cloud computing. IEEE Trans Parallel Distrib Syst 33(6):1503–1519
Alhelaly S, Muthanna A, Elgendy IA (2022) Optimizing task offloading energy in multi-user multi-uav-enabled mobile edge-cloud computing systems. Appl Sci 12(13):6566
Cozzolino V, Tonetto L, Mohan N, Ding AY, Ott J (2022) Nimbus: Towards latency-energy efficient task offloading for ar services. IEEE Trans Cloud Comput
Jin X, Hua W,Wang Z, Chen Y (2022) A survey of research on computation offloading in mobile cloud computing. Wireless Networks 1–23
Hu S, Xiao Y (2021) Design of cloud computing task offloading algorithm based on dynamic multi-objective evolution. Future Generation Computer Systems 122:144–148
De D, Mukherjee A, Guha Roy D (2020) Power and delay efficient multilevel offloading strategies for mobile cloud computing. Wireless Personal Communications 112(4):2159–2186
Plachy J, Becvar Z, Strinati EC, Pietro Nd (2021) Dynamic allocation of computing and communication resources in multi-access edge computing for mobile users. IEEE Trans Netw Serv Manag 18(2):2089–2106. https://doi.org/10.1109/TNSM.2021.3072433
uz Zaman SK, Jehangiri AI, Maqsood T, Ahmad Z, Umar AI, Shuja J, Alanazi E, Alasmary W (2021) Mobility-aware computational offloading in mobile edge networks: a survey. Cluster Computing 1–22
Chakraborty S, De D, Mazumdar K (2022) Dome: Dew computing based microservice execution in mobile edge using q-learning. Appl Intell 1–20
Shuja J, Bilal K, Alasmary W, Sinky H, Alanazi E (2021) Applying machine learning techniques for caching in next-generation edge networks: A comprehensive survey. J Netw Comput App 103005
Zhao F, Chen Y, Zhang Y, Liu Z, Chen X (2021) Dynamic offloading and resource scheduling for mobile-edge computing with energy harvesting devices. IEEE Trans Netw Serv Manag 18(2):2154–2165. https://doi.org/10.1109/TNSM.2021.3069993
Tian K, Chai H, Liu Y, Liu B (2022) Edge intelligence empowered dynamic offloading and resource management of mec for smart city internet of things. Electronics 11(6):879
Gao M, Shen R, Li J, Yan S, Li Y, Shi J, Han Z, Zhuo L (2020) Computation offloading with instantaneous load billing for mobile edge computing. IEEE Trans Serv Comput
Alfakih T, Hassan MM, Gumaei A, Savaglio C, Fortino G (2020) Task offloading and resource allocation for mobile edge computing by deep reinforcement learning based on sarsa. IEEE Access 8:54074–54084
Hadi M, Ghazizadeh R (2022) Joint resource allocation, user clustering and 3-d location optimization in multi-uav-enabled mobile edge computing. Computer Networks 109420
Wang Z, Lv T, Chang Z (2022) Computation offloading and resource allocation based on distributed deep learning and software defined mobile edge computing. Computer Networks 205:108732
Lu W, Mo Y, Feng Y, Gao Y, Zhao N, Wu Y, Nallanathan A (2022) Secure transmission for multi-uav-assisted mobile edge computing based on reinforcement learning. IEEE Trans Netw Sci Eng
Jitani A, Mahajan A, Zhu Z, Abou-Zeid H, Fapi ET, Purmehdi H (2022) Structure-aware reinforcement learning for node-overload protection in mobile edge computing. IEEE Trans Cogn Commun Netw
Sutton RS, Barto AG et al (1998) Introduction to Reinforcement Learning vol 135. MIT press Cambridge, ???
Sutton RS, McAllester DA, Singh SP, Mansour Y (2000) Policy gradient methods for reinforcement learning with function approximation. In: Advances in Neural Information Processing Systems 1057–1063
Lyu Y, Liu Z, Fan R, Zhan C, Hu H, An J (2022) Optimal computation offloading in collaborative leo-iot enabled mec: A multi-agent deep reinforcement learning approach. IEEE Trans Green Commun Netw
Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, Graves A, Riedmiller M, Fidjeland AK, Ostrovski G et al (2015) Human-level control through deep reinforcement learning. Nature 518(7540):529
Van Hasselt H, Guez A, Silver D (2016) Deep reinforcement learning with double q-learning. In: Thirtieth AAAI Conference on Artificial Intelligence
Wang Z, Schaul T, Hessel M, Hasselt H, Lanctot M, Freitas N (2016) Dueling network architectures for deep reinforcement learning. In: International Conference on Machine Learning 1995–2003
Kakade SM (2001) A natural policy gradient. Advances in neural information processing systems 14
Mnih V, Badia AP, Mirza M, Graves A, Lillicrap T, Harley T, Silver D, Kavukcuoglu K (2016) Asynchronous methods for deep reinforcement learning. In: International Conference on Machine Learning 1928–1937
Haarnoja T, Zhou A, Abbeel P, Levine S (2018) Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor. In: International Conference on Machine Learning 1861–1870 PMLR
Lillicrap TP, Hunt JJ, Pritzel A, Heess N, Erez T, Tassa Y, Silver D, Wierstra D (2016) Continuous control with deep reinforcement learning 1–14
Schulman J, Wolski F, Dhariwal P, Radford A, Klimov O (2017) Proximal policy optimization algorithms. https://doi.org/https://arxiv.org/pdf/1707.06347.pdf
Abouaomar A, Mlika Z, Filali A, Cherkaoui S, Kobbane A (2021) A deep reinforcement learning approach for service migration in mec-enabled vehicular networks. In: 2021 IEEE 46th Conference on Local Computer Networks (LCN) 273–280. IEEE
Wang L, Wang K, Pan C, Xu W, Aslam N, Nallanathan A (2021) Deep reinforcement learning based dynamic trajectory control for uav-assisted mobile edge computing. IEEE Transactions on Mobile Computing
Karimi E, Chen Y, Akbari B (2022) Task offloading in vehicular edge computing networks via deep reinforcement learning. Computer Communications 189:193–204
Nduwayezu M, Yun JH (2022) Latency and energy aware rate maximization in mc-noma-based multi-access edge computing: A two-stage deep reinforcement learning approach. Computer Networks 207:108834
Ngo HQ, Larsson EG, Marzetta TL (2013) Energy and spectral efficiency of very large multiuser mimo systems. IEEE Trans Commun 61(4):1436–1449
Ke H, Wang J, Deng L, Ge Y, Wang H (2020) Deep reinforcement learning-based adaptive computation offloading for mec in heterogeneous vehicular networks. IEEE Trans Veh Technol 69(7):7916–7929
Chen Z, Zhang L, Pei Y, Jiang C, Yin L (2021) Noma-based multi-user mobile edge computation offloading via cooperative multi-agent deep reinforcement learning. IEEE Trans Cogn Commun Netw 8(1):350–364
Chen Z (2020) Wang X (2020) Decentralized computation offloading for multiuser mobile edge computing: A deep reinforcement learning approach. EURASIP Journal on Wireless Communications and Networking 1:1–21
Kuang Z, Shi Y, Guo S, Dan J, Xiao B (2019) Multi-user offloading game strategy in ofdma mobile cloud computing system. IEEE Trans Veh Technol 68(12):12190–12201
Wu Y, Wang Y, Zhou F, Hu RQ (2019) Computation efficiency maximization in ofdma-based mobile edge computing networks. IEEE Commun Lett 24(1):159–163
Chen X, Jiao L, Li W, Fu X (2015) Efficient multi-user computation offloading for mobile-edge cloud computing. IEEE/ACM transactions on networking 24(5):2795–2808
Acknowledgements
This work was supported by the Jilin Provincial Science and Technology Department Natural Science Foundation of China (20210101415JC) and the Jilin Provincial Science and Technology Department Free Exploration Research Project of China (YDZJ202201ZYTS556).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare that there are no conflicts of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Hongchang Ke, Hui Wang and Hongbin Sun These authors contributed equally to this work.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Ke, H., Wang, H. & Sun, H. Deep Reinforcement Learning-based Power Control and Bandwidth Allocation Policy for Weighted Cost Minimization in Wireless Networks. Appl Intell 53, 26885–26906 (2023). https://doi.org/10.1007/s10489-023-04929-2
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-04929-2