10.57647/j.mjee.2025.17785

Optimizing Network Performance in Mobile Data Offloading Using Cooperative IEEE 802.11 MAC Protocol and Deep Reinforcement Learning

  1. Institute of Artificial Intelligence and Social and Advanced Technologies, Isf.C., Islamic Azad University, Isfahan, Iran
  2. Department of Electrical Engineering, DOL.C., Islamic Azad University, Isfahan, Iran
  3. Department of Electronic Engineering, Southern Technical University, Amara, Iraq

Published Online: 2025-10-21

How to Cite

Alrashedi, N. A. Y., Sadeghi, R., Al-Lamy, W. H. Z., Hamidkhani, M., & Khorsand, R. Optimizing Network Performance in Mobile Data Offloading Using Cooperative IEEE 802.11 MAC Protocol and Deep Reinforcement Learning. Majlesi Journal of Electrical Engineering. https://doi.org/10.57647/j.mjee.2025.17785

PDF views: 136

Abstract

With the exponential growth of mobile data traffic and the increasing demand for latency-sensitive applications, wireless networks face critical resource management challenges. In particular, heterogeneous architectures such as LTE and Wi-Fi necessitate intelligent and adaptive offloading mechanisms to ensure efficiency and Quality of Service (QoS). This paper exploits Deep Reinforcement Learning (DRL) and cooperative communication within the MAC layer of the IEEE 802.11n standard to design an adaptive and scalable framework for data offloading in wireless networks. The proposed model employs the Partially Observable Markov Decision Process (POMDP) structure to optimally select the communication path and the type of communication (direct or multi-hop) in real-time. The proposed learning algorithm, which leverages the Deep Q-Learning structure and the Policy Improvement mechanism, demonstrates significant performance gains compared to conventional methods. Simulation results indicate that the average cumulative reward achieved by the DRL algorithm is 3.4, a substantial improvement in decision-making effectiveness compared to 2.2 for Q-Learning and 0.7 for Heuristic methods. Moreover, the energy efficiency of the proposed method improved by 87% compared to CoopMAC, and the throughput increased by 18%. These results establish the proposed framework as an effective and low-power solution for implementation in Fifth-Generation (5G) network architectures.

Keywords

  • Mobile data offloading,
  • Deep reinforcement learning,
  • Cooperative relay selection,
  • CoopMAC,
  • Real-time optimization,
  • Q-learning

References

  1. R. Yan, Z. Guo, P. Liu, Q. Lan, X.-P. Zhang, and Y. Dong, “Multi-Agent Reinforcement Learning Based Channel Access Optimization for IEEE 802.11 bn,” IEEE Transactions on Green Communications and Networking, 2024. DOI: https://doi.org/ 10.1109/TGCN.2024.3495236
  2. Y. Yang, and S. Yan, “Joint Throughput Maximization and Energy Management for Ultra-low Power Ambient Backscatter Communication in WBANs by Distributed Deep Reinforcement Learning,” IEEE Sensors Journal, 2024. DOI: https://doi.org/10.1109/JSEN.2024.3487354.
  3. A. A. Chirani, “Network reconfiguration and optimal distributed generations allocation with whale optimizer algorithm,” Majlesi Journal of Electrical Engineering, vol. 19, no. 1 (March 2025), pp. 1-12, 2025. DOI: https://doi.org/10.57647/j.mjee.2025.1901.04
  4. S. Song, Z. Zhang, Q. Wu, P. Fan, and Q. Fan, “Joint optimization of age of information and energy consumption in NR-V2X system based on deep reinforcement learning,” Sensors, vol. 24, no. 13, pp. 4338, 2024, DOI: https://doi.org/10.3390/s24134338.
  5. M. Alizadeh Aliabadi, M. Karimi, Z. Karimi, and M. soheili Fard, “The Effect of Photoplethysmography Signal Denoising on Compression Quality,” IRANIAN JOURNAL OF ELECTRICAL AND ELECTRONIC ENGINEERING, vol. 21, no. 1, pp. 3277-3277, 2025, DOI: https://doi.org/10.22068/IJEEE.21.1.3277
  6. M. J. Kadhim, R. Sadeghi, A. S. Abdalrada, B. Arandian, and R. Khorsand, “Performance improvement of data offloading using Krill herd optimization algorithm,” Majlesi Journal of Electrical Engineering, vol. 19, no. 1 (March 2025), pp. 17-17, 2025, DOI: https://doi.org/10.57647/j.mjee.2025.1901.05
  7. A. Khoshnoudi, R. Sadeghi, and F. Faghani, “Performance Improvement of Data Offloading using Multi-rate IEEE 802.11 WLAN,” Majlesi Journal of Electrical Engineering, vol. 13, no. 1, pp. 121-126, 2019.
  8. N. Dawar, K. N. Nguyen, A. Sehgal, Y. Zhu, B. L. Ng, and J. Choi, “Enhancing Wi-Fi 7: Traffic Flow Intelligence and Multi-Link Operation for Optimal Efficiency,” IEEE Access, 2025, DOI: https://doi.org/10.1109/ACCESS.2025.3557435
  9. A. A. Alaidany, and M. M. Mahdi, “A Review of IoT-Based Wearable Sensor Systems for Healthcare Monitoring., DOI:No DOI
  10. M. Talebkhah, A. Sali, V. Khodamoradi, T. Khodadadi, and M. Gordan, “Task offloading for edge-IoV networks in the industry 4.0 era and beyond: a high-level view,” Engineering Science and Technology, an International Journal, vol. 54, pp. 101699, 2024, DOI: https://doi.org/10.1016/j.jestch.2024.101699
  11. Z. Zabihi, A. M. Eftekhari Moghadam, and M. H. Rezvani, “Reinforcement learning methods for computation offloading: a systematic review,” ACM Computing Surveys, vol. 56, no. 1, pp. 1-41, 2023, DOI: https://doi.org/10.1145/3603703
  12. M. Harouni, M. Karimi, A. Nasr, H. Mahmoudi, and Z. Arab Najafabadi, "Health monitoring methods in heart diseases based on data mining approach: A directional review," Prognostic models in healthcare: Ai and statistical approaches, pp. 115-159: Springer, 2022, DOI: https://doi.org/10.1007/978-981-19-2057-8_5
  13. E. T. Garmaserh, and M. Emadi, “Improving the criteria of electricity consumptionforecasting in petrochemical industrial units based ondeep learning,” Majlesi Journal of Electrical Engineering, vol. 19, no. 2 (June 2025), 2025, DOI: https://doi.org/10.57647/j.mjee.2025.1902.41
  14. S. S. S. Abolghasemi, M. Emadi, and M. Karimi, “Accuracy improvement of breast tumor detection based on dimension reduction in the spatial and edge features and edge structure in the image,” Majlesi Journal of Electrical Engineering, vol. 18, no. 1, 2024, DOI: https://doi.org/10.30486/mjee.2023.1991110.1174
  15. B. Kar, W. Yahya, Y.-D. Lin, and A. Ali, “Offloading using traditional optimization and machine learning in federated cloud–edge–fog systems: A survey,” IEEE Communications Surveys & Tutorials, vol. 25, no. 2, pp. 1199-1226, 2023, DOI: https://doi.org/ 10.1109/COMST.2023.3239579
  16. S. Dong, J. Tang, K. Abbas, R. Hou, J. Kamruzzaman, L. Rutkowski, and R. Buyya, “Task offloading strategies for mobile edge computing: A survey,” Computer Networks, pp. 110791, 2024, DOI: https://doi.org/10.1016/j.comnet.2024.110791
  17. R. Chaari, O. Cheikhrouhou, A. Koubâa, H. Youssef, and T. N. Gia, “Dynamic computation offloading for ground and flying robots: Taxonomy, state of art, and future directions,” Computer Science Review, vol. 45, pp. 100488, 2022, DOI: https://doi.org/ 10.1016/j.cosrev.2022.100488
  18. Y. Li, G. Su, P. Hui, D. Jin, L. Su, and L. Zeng, "Multiple mobile data offloading through delay tolerant networks." pp. 43-48., 2011, DOI: https://doi.org/10.1145/2030652.2030665
  19. S. Andreev, A. Pyattaev, K. Johnsson, O. Galinina, and Y. Koucheryavy, “Cellular traffic offloading onto network-assisted device-to-device connections,” IEEE Communications Magazine, vol. 52, no. 4, pp. 20-31, 2014, DOI: https://doi.org/10.1109/MCOM.2014.6807943
  20. K. Lee, J. Lee, Y. Yi, I. Rhee, and S. Chong, “Mobile data offloading: How much can WiFi deliver?,” IEEE/ACM Transactions on networking, vol. 21, no. 2, pp. 536-550, 2012, DOI: https://doi.org/10.1109/TNET.2012.2218122
  21. C. P. Mayer, and O. P. Waldhorst, "Offloading infrastructure using delay tolerant networks and assurance of delivery." pp. 1-7, 2011, DOI: https://doi.org/10.1109/WirelessDays.2011.6134105
  22. X. Wang, M. Chen, Z. Han, D. O. Wu, and T. T. Kwon, "TOSS: Traffic offloading by social network service-based opportunistic sharing in mobile social networks." pp. 2346-2354, DOI: https://doi.org/10.1109/INFOCOM.2014.6848179
  23. A. Anagnostopoulos, R. Kumar, and M. Mahdian, "Influence and correlation in social networks." pp. 7-15 ,2008, DOI: https://doi.org/10.1145/1401890.1401897
  24. J. Y. Ryu, J. Lee, and T. Q. Quek, “Confidential cooperative communication with trust degree of potential eavesdroppers,” IEEE Transactions on Wireless Communications, vol. 15, no. 6, pp. 3823-3836, 2016, DOI: https://doi.org/ 10.1109/TWC.2016.2530058
  25. N. Magaia, Z. Sheng, P. R. Pereira, and M. Correia, “REPSYS: A robust and distributed incentive scheme for collaborative caching and dissemination in content-centric cellular-based vehicular delay-tolerant networks,” IEEE Wireless Communications, vol. 25, no. 3, pp. 65-71, 2018, DOI: https://doi.org/10.1109/MWC.2018.1700284
  26. D. Huang, P. Wang, and D. Niyato, “A dynamic offloading algorithm for mobile computing,” IEEE Transactions on Wireless Communications, vol. 11, no. 6, pp. 1991-1995, 2012, DOI: https://doi.org/10.1109/TWC.2012.041912.110912
  27. L. Valerio, R. Bruno, and A. Passarella, "Adaptive data offloading in opportunistic networks through an actor-critic learning method." pp. 31-36, DOI: https://doi.org/10.1145/2645672.2645676
  28. F. Rebecchi, M. D. De Amorim, V. Conan, A. Passarella, R. Bruno, and M. Conti, “Data offloading techniques in cellular networks: A survey,” IEEE Communications Surveys & Tutorials, vol. 17, no. 2, pp. 580-603, 2014, DOI: https://doi.org/10.1109/COMST.2014.2369742
  29. G. Gao, M. Xiao, J. Wu, K. Han, and L. Huang, "Deadline-sensitive mobile data offloading via opportunistic communications." pp. 1-9, 2017, DOI: https://doi.org/10.1109/TPDS.2017.2720741
  30. P. Hui, A. Chaintreau, J. Scott, R. Gass, J. Crowcroft, and C. Diot, "Pocket switched networks and human mobility in conference environments." pp. 244-251, DOI: https://doi.org/10.1145/1080139.1080142
  31. S. G, K. K, and V. M, “A novel task offloading model for IoT: enhancing resource utilization with actor-critic-based reinforcement learning,” Earth Science Informatics, vol. 18, no. 3, pp. 266, 2025/02/17, 2025, DOI: https://doi.org/ 10.1007/s12145-025-01773-5
  32. C. Liu, H. Wang, M. Zhao, J. Liu, X. Zhao, and P. Yuan, “Dependency-aware online task offloading based on deep reinforcement learning for IoV,” Journal of Cloud Computing, vol. 13, no. 1, pp. 136, 2024, DOI: https://doi.org/ 10.1186/s13677-024-00701-0