Abstract
Efficient energy utilization in wireless sensor networks is paramount for the success of Internet of things (IoT) applications. Traditional energy-saving routing protocols typically compute isolated routes, which may only partially optimize energy consumption globally. Software-defined wireless sensor networking (SDWSN) is emerging as a crucial architecture for enabling IoT, leveraging software-defined network principles to facilitate routing operations. However, existing routing algorithms often need more capability to identify the most optimized paths. In this paper, we propose a reinforcement learning-based energy-aware routing (RL-EAR) algorithm to optimize the routing paths of SDWSN. Our method utilizes a reward function considering factors such as remaining energy, hop count, and congestion to optimize routes from any node to the controller node, aiming to enhance network lifetime. The SDWSN controller, acting as the agent, receives rewards based on its actions and refines routes over time through interactions with the data plane. The simulation results of the RL-EAR algorithm demonstrate its prominence over existing routing algorithms. Metrics such as network lifetime, packet delivery ratio, and the number of dead nodes over time are utilized to evaluate the effectiveness of our proposed algorithm. As a result, RL-EAR increases network lifetime by 42% compared to existing routing algorithms.










Similar content being viewed by others
Data availability
No datasets were generated or analyzed during the current study.
References
Kandris D, Nakas C, Vomvas D, Koulouras G (2020) Applications of wireless sensor networks: an up-to-date survey. Appl Syst Innov 3(1):14
Nunes BAA, Mendonca M, Nguyen X-N, Obraczka K, Turletti T (2014) A survey of software-defined networking: past, present, and future of programmable networks. IEEE Commun Surv Tutor 16(3):1617–1634
Narwaria A, Mazumdar AP (2023) Software-defined wireless sensor network: a comprehensive survey. J Netw Comput Appl 215:103636
Narwaria A, Mazumdar AP, Sharma S (2023) Energy efficient multi-objective task allocation in software-defined wireless sensor network. In: 2023 IEEE 8th International Conference for Convergence in Technology (I2CT), pp 1–6
Rahimifar A, Kavian YS, Kaabi H, Soroosh M (2023) A smart duty cycle for lifetime enhancement and control overhead in SDWSN. Iran J Sci Technol Trans Electr Eng 47(3):1207–1223
Moerland TM, Broekens J, Plaat A, Jonker CM et al (2023) Model-based reinforcement learning: a survey. Found Trends® Mach Learn 16(1):1–118
Savaglio C, Pace P, Aloi G, Liotta A, Fortino G (2019) Lightweight reinforcement learning for energy efficient communications in wireless sensor networks. IEEE Access 7:29355–29364
Rahimifar A, Seifi Kavian Y, Kaabi H, Soroosh M (2021) Predicting the energy consumption in software defined wireless sensor networks: a probabilistic Markov model approach. J Ambient Intell Humaniz Comput 12(10):9053–9066
Younus MU, Khan MK, Bhatti AR (2021) Improving the software-defined wireless sensor networks routing performance using reinforcement learning. IEEE Internet Things J 9(5):3495–3508
Kobo HI, Abu-Mahfouz AM, Hancke GP (2017) A survey on software-defined wireless sensor networks: challenges and design requirements. IEEE Access 5:1872–1899
Vijayan K, Kshirsagar PR, Sonekar SV, Chakrabarti P, Unhelkar B, Margala M (2024) Optimizing IoT-enabled WSN routing strategies using whale optimization-driven multi-criterion correlation approach employs the reinforcement learning agent. Opt Quantum Electron 56(4):568
Priyadarshi R (2024) Exploring machine learning solutions for overcoming challenges in IoT-based wireless sensor network routing: a comprehensive review. Wirel Netw 30:1–27
Jurado-Lasso FF, Clarke K, Cadavid AN, Nirmalathas A (2022) Energy-aware routing for software-defined multihop wireless sensor networks. IEEE Sens J 21(8):10174–10182
Yang X, Yan J, Wang D, Xu Y, Hua G (2024) WOAD3QN-RP: an intelligent routing protocol in wireless sensor networks-a swarm intelligence and deep reinforcement learning based approach. Expert Syst Appl 246:123089
Garcia F, Rachelson E (2013) Markov decision processes. Markov Decision Processes in Artificial Intelligence, pp 1–38
Xiang W, Wang N, Zhou Y (2016) An energy-efficient routing algorithm for software-defined wireless sensor networks. IEEE Sens J 16(20):7393–7400
Kamarei M, Tavakoli F, Patooghy A (2022) Route packets uneven, consume energy even: A lifetime expanding routing algorithm for WSNS. Wireless Pers Commun 125(4):3133–3151
Kamarei M, Patooghy A, Alsharif A, Hakami V (2020) Simple: a unified single and multi-path routing algorithm for wireless sensor networks with source location privacy. IEEE Access 8:33818–33829
Wenxing L, Muqing W, Yuewei W (2016) Energy-efficient algorithm based on multi-dimensional energy space for software-defined wireless sensor networks. In: 2016 International Symposium on Wireless Communication Systems (ISWCS), IEEE, pp 309–314
Kim S, Son J, Talukder A, Hong CS (2016) Congestion prevention mechanism based on q-leaning for efficient routing in SDN. In: 2016 International Conference on Information Networking (ICOIN), IEEE, pp 124–128
Stampa G, Arias M, Sánchez-Charles D, Muntés-Mulero V, Cabellos A (2017) A deep-reinforcement learning approach for software-defined networking routing optimization. arXiv:1709.07080
Mostafaei H, Obaidat MS (2018) Learning automaton-based self-protection algorithm for wireless sensor networks. IET Netw 7(5):353–361
Guo W, Yan C, Lu T (2019) Optimizing the lifetime of wireless sensor networks via reinforcement-learning-based routing. Int J Distrib Sens Netw 15(2):1550147719833541
Lee S, Seo Y, Lee K, Abbeel P, Shin J (2022) Offline-to-online reinforcement learning via balanced replay and pessimistic q-ensemble. In: Conference on Robot Learning, PMLR, pp 1702–1712
Fang C, Cheng C, Tang Z, Li C (2019) Research on routing algorithm based on reinforcement learning in SDN. In: Journal of Physics: Conference Series, vol 1284. IOP Publishing, p 012053
Younus MU, Khan MK, Anjum MR, Afridi S, Arain ZA, Jamali AA (2020) Optimizing the lifetime of software defined wireless sensor network via reinforcement learning. IEEE Access 9:259–272
Jurado-Lasso FF, Clarke K, Cadavid AN, Nirmalathas A (2021) Energy-aware routing for software-defined multihop wireless sensor networks. IEEE Sens J 21(8):10174–10182
Dunkels A, Gronvall B, Voigt T (2004) Contiki-a lightweight and flexible operating system for tiny networked sensors. In: 29th Annual IEEE International Conference on Local Computer Networks, IEEE, pp 455–462
Eriksson J, Österlind F, Finne N, Tsiftes N, Dunkels A, Voigt T, Sauter R, Marrón PJ (2009) Cooja/mspsim: interoperability testing for wireless sensor networks. In: Proceedings of the 2nd International Conference on Simulation Tools and Techniques, pp 1–7
Kumar A, Zhou A, Tucker G, Levine S (2020) Conservative q-learning for offline reinforcement learning. Adv Neural Inf Process Syst 33:1179–1191
Casas-Velasco DM, Rendon OMC, Fonseca NL (2020) Intelligent routing based on reinforcement learning for software-defined networking. IEEE Trans Netw Serv Manage 18(1):870–881
Narwaria A, Mazumdar AP, Kalla G (2021) C3hac: A controller placement approach for sdwsn. In: TENCON 2021-2021 IEEE Region 10 Conference (TENCON), IEEE, pp 935–940
Knox WB, Stone P (2012) Reinforcement learning from simultaneous human and MDP reward. In: AAMAS, vol 1004. Valencia, pp 475–482
Su X, Ren Y, Cai Z, Liang Y, Guo L (2022) A Q-learning based routing approach for energy efficient information transmission in wireless sensor network. IEEE Trans Netw Serv Manag 20(2):1949–1961
Van Seijen H, Fatemi M, Romoff J, Laroche R, Barnes T, Tsang J (2017) Hybrid reward architecture for reinforcement learning. Advances in Neural Information Processing Systems 30
Narwaria A, Soni K, Mazumdar AP (2024) A position and energy aware multi-objective controller placement and re-placement scheme in distributed SDWSN. J Supercomput 80:1–29
Author information
Authors and Affiliations
Contributions
A. N. contributed to the conceptualization, methodology, writing, reviewing, editing, and software development of the study. V. K. was responsible for methodology, software implementation, and the creation of figures and graphs. A. P. M. contributed to conceptualization, reviewing, editing, and provided supervision for the research.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Narwaria, A., Kumari, V. & Mazumdar, A.P. RL-EAR: reinforcement learning-based energy-aware routing for software-defined wireless sensor network. J Supercomput 81, 485 (2025). https://doi.org/10.1007/s11227-025-06998-1
Accepted:
Published:
DOI: https://doi.org/10.1007/s11227-025-06998-1