skip to main content
10.1145/3345768.3355908acmconferencesArticle/Chapter ViewAbstractPublication PagesmswimConference Proceedingsconference-collections
research-article

ns-3 meets OpenAI Gym: The Playground for Machine Learning in Networking Research

Published: 25 November 2019 Publication History

Abstract

Recently, we have seen a boom of attempts to improve the operation of networking protocols using machine learning techniques. The proposed reinforcement learning (RL) based control solutions very often overtake traditionally designed ones in terms of performance and efficiency. However, in order to reach such a superb level, an RL control agent requires a lot of interactions with an environment to learn the best policies. Similarly, the recent advancements in image recognition area were enabled by the rise of large labeled datasets (e.g. ImageNet). This paper presents the ns3-gym - the first framework for RL research in networking. It is based on OpenAI Gym, a toolkit for RL research and ns-3 network simulator. Specifically, it allows representing an ns-3 simulation as an environment in Gym framework and exposing state and control knobs of entities from the simulation for the agent's learning purposes. Our framework is generic and can be used in various networking problems. Here, we present an illustrative example from the cognitive radio area, where a wireless node learns the channel access pattern of a periodic interferer in order to avoid collisions with it. The toolkit is provided to the community as open-source under a GPL license.

References

[1]
Dafna Shahaf Aviv Tamar Asaf Valadarsky, Michael Schapira. 2017. Learning To Route with Deep RL. In NIPS.
[2]
R. Atallah, C. Assi, and M. Khabbaz. 2017. Deep reinforcement learning-based scheduling for roadside communication networks. In WiOpt.
[3]
Sergio Barrachina-Muñoz, Francesc Wilhelmi, Ioannis Selinis, and Boris Bellalta. 2018. Komondor: a Wireless Network Simulator for Next-Generation High- Density WLANs. CoRR abs/1811.12397 (2018). arXiv:1811.12397 http://arxiv.org/ abs/1811.12397
[4]
Greg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. 2016. OpenAI Gym. CoRR (2016). http://arxiv.org/abs/1606.01540
[5]
Gustavo Carneiro, Helder Fontes, and Manuel Ricardo. 2011. Fast prototyping of network protocols through ns-3 simulation model reuse. Simulation modelling practice and theory, Elsevier (2011).
[6]
Sandeep Chinchali, Pan Hu, Tianshu Chu, Manu Sharma, Manu Bansal, Rakesh Misra, Marco Pavone, and Sachin Katti. 2018. Cellular Network Traffic Scheduling With Deep Reinforcement Learning. In AAAI.
[7]
Paul F. Christiano, Zain Shah, Igor Mordatch, Jonas Schneider, Trevor Blackwell, Joshua Tobin, Pieter Abbeel, and Wojciech Zaremba. 2016. Transfer from Simulation to Real World through Learning Deep Inverse Dynamics Model. CoRR (2016). http://arxiv.org/abs/1610.03518
[8]
J. Deng, W. Dong, R. Socher, L. Li, Kai Li, and Li Fei-Fei. 2009. ImageNet: A large-scale hierarchical image database. In IEEE CVPR.
[9]
Prafulla Dhariwal, Christopher Hesse, Oleg Klimov, Alex Nichol, Matthias Plappert, Alec Radford, John Schulman, Szymon Sidor, YuhuaiWu, and Peter Zhokhov. 2017. OpenAI Baselines. https://github.com/openai/baselines.
[10]
Yan Duan, Xi Chen, Rein Houthooft, John Schulman, and Pieter Abbeel. 2016. Benchmarking Deep Reinforcement Learning for Continuous Control. CoRR abs/1604.06778 (2016). arXiv:1604.06778 http://arxiv.org/abs/1604.06778
[11]
Rohit Gupta, Bjoern Bachmann, Russell Ford, Sundeep Rangan, Nikhil Kundargi, Amal Ekbal, Karamvir Rathi, Maria Isabel Sanchez, Antonio de la Oliva, and Arianna Morelli. 2015. Ns-3-based Real-time Emulation of LTE Testbed Using LabVIEW Platform for Software Defined Networking (SDN) in CROWD Project. In Proceedings of the 2015 Workshop on Ns-3 (WNS3 '15).
[12]
Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. In ICLR. http://arxiv.org/abs/1412.6980
[13]
Jens Kober, J Andrew Bagnell, and Jan Peters. 2013. Reinforcement learning in robotics: A survey. The International Journal of Robotics Research (2013).
[14]
Yiming Kong, Hui Zang, and Xiaoli Ma. 2018. Improving TCP Congestion Control with Machine Intelligence. In ACM NetAI.
[15]
Wei Li, Fan Zhou, Kaushik Roy Chowdhury, and Waleed M Meleis. 2018. QTCP: Adaptive Congestion Control with Reinforcement Learning. IEEE Transactions on Network Science and Engineering (2018).
[16]
Eric Liang, Richard Liaw, Robert Nishihara, Philipp Moritz, Roy Fox, Ken Goldberg, Joseph E. Gonzalez, Michael I. Jordan, and Ion Stoica. 2018. RLlib: Abstractions for Distributed Reinforcement Learning. In International Conference on Machine Learning (ICML).
[17]
Ryan Lowe, Yi Wu, Aviv Tamar, Jean Harb, OpenAI Pieter Abbeel, and Igor Mordatch. 2017. Multi-agent actor-critic for mixed cooperative-competitive environments. In Advances in Neural Information Processing Systems.
[18]
Hongzi Mao, Mohammad Alizadeh, Ishai Menache, and Srikanth Kandula. 2016. Resource Management with Deep Reinforcement Learning. In ACM HotNets.
[19]
Hongzi Mao, Ravi Netravali, and Mohammad Alizadeh. 2017. Neural adaptive video streaming with pensieve. In ACM SIGCOMM.
[20]
Volodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Timothy P. Lillicrap, Tim Harley, David Silver, and Koray Kavukcuoglu. 2016. Asynchronous Methods for Deep Reinforcement Learning. CoRR (2016). http: //arxiv.org/abs/1602.01783
[21]
NS-3 Consortium. [n.d.]. ns-3 documentation. https:www.nsnam.org. Accessed: 2019-07--20.
[22]
NS-3 Consortium. [n.d.]. ns-3 source code. http:code.nsnam.org. Accessed: 2019-07--20.
[23]
OpenAI. [n.d.]. OpenAI Gym documentation. https:gym.openai.com. Accessed: 2019-07--20.
[24]
OpenAI. [n.d.]. OpenAI Gym source code. https:github.comopenaigym. Accessed: 2019-07--20.
[25]
Wojciech Samek, Slawomir Stanczak, and Thomas Wiegand. 2017. The Convergence of Machine Learning and Communications. CoRR (2017). http: //arxiv.org/abs/1708.08299
[26]
P. Lalith Suresh and Ruben Merz. 2011. Ns-3-click: Click Modular Router Integration for Ns-3. In Proceedings of the 4th International ICST Conference on Simulation Tools and Techniques (SIMUTools '11). 8.
[27]
Hajime Tazaki, Frédéric Uarbani, Emilio Mancini, Mathieu Lacage, Daniel Camara, Thierry Turletti, and Walid Dabbous. 2013. Direct Code Execution: Revisiting Library OS Architecture for Reproducible Network Experiments. In ACM CoNEXT.
[28]
M. Wang, Y. Cui, X. Wang, S. Xiao, and J. Jiang. 2018. Machine Learning for Networking: Workflow, Advances and Opportunities. IEEE Network (2018).
[29]
Keith Winstein and Hari Balakrishnan. 2013. TCP Ex Machina: Computergenerated Congestion Control. In ACM SIGCOMM.
[30]
Iker Zamora, Nestor Gonzalez Lopez, Victor Mayoral Vilches, and Alejandro Hernandez Cordero. 2016. Extending the OpenAI Gym for robotics: a toolkit for reinforcement learning using ROS and Gazebo. arXiv preprint arXiv:1608.05742 (2016).

Cited By

View all
  • (2024)Enabling Online Reinforcement Learning Training for Open RAN2024 IFIP Networking Conference (IFIP Networking)10.23919/IFIPNetworking62109.2024.10619796(577-582)Online publication date: 3-Jun-2024
  • (2024)Wixor: Dynamic TDD Policy Adaptation for 5G/xG NetworksProceedings of the ACM on Networking10.1145/36963952:CoNEXT4(1-24)Online publication date: 25-Nov-2024
  • (2024)RayNet: A Simulation Platform for Developing Reinforcement Learning-Driven Network ProtocolsACM Transactions on Modeling and Computer Simulation10.1145/365397534:3(1-25)Online publication date: 30-Mar-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MSWIM '19: Proceedings of the 22nd International ACM Conference on Modeling, Analysis and Simulation of Wireless and Mobile Systems
November 2019
340 pages
ISBN:9781450369046
DOI:10.1145/3345768
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 25 November 2019

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. network simulator
  2. networking research
  3. ns-3
  4. openai gym
  5. reinforcement learning

Qualifiers

  • Research-article

Funding Sources

  • Bundesministeium für Wirtschaft und Energie

Conference

MSWiM '19
Sponsor:

Acceptance Rates

Overall Acceptance Rate 398 of 1,577 submissions, 25%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)414
  • Downloads (Last 6 weeks)47
Reflects downloads up to 03 Mar 2025

Other Metrics

Citations

Cited By

View all
  • (2024)Enabling Online Reinforcement Learning Training for Open RAN2024 IFIP Networking Conference (IFIP Networking)10.23919/IFIPNetworking62109.2024.10619796(577-582)Online publication date: 3-Jun-2024
  • (2024)Wixor: Dynamic TDD Policy Adaptation for 5G/xG NetworksProceedings of the ACM on Networking10.1145/36963952:CoNEXT4(1-24)Online publication date: 25-Nov-2024
  • (2024)RayNet: A Simulation Platform for Developing Reinforcement Learning-Driven Network ProtocolsACM Transactions on Modeling and Computer Simulation10.1145/365397534:3(1-25)Online publication date: 30-Mar-2024
  • (2024)5G Real-Time QoS-Driven Packet Scheduler for O-RAN2024 IEEE 99th Vehicular Technology Conference (VTC2024-Spring)10.1109/VTC2024-Spring62846.2024.10683043(1-6)Online publication date: 24-Jun-2024
  • (2024)On Learning Suitable Caching Policies for In-Network CachingIEEE Transactions on Machine Learning in Communications and Networking10.1109/TMLCN.2024.34364722(1076-1092)Online publication date: 2024
  • (2024)Cooperate or Not Cooperate: Transfer Learning With Multi-Armed Bandit for Spatial Reuse in Wi-FiIEEE Transactions on Machine Learning in Communications and Networking10.1109/TMLCN.2024.33719292(351-369)Online publication date: 2024
  • (2024)Multiple QoS Enabled Intelligent Resource Management in Vehicle-to-Vehicle CommunicationIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2024.336555725:9(12081-12094)Online publication date: Sep-2024
  • (2024)Deep Reinforcement Learning Based Dynamic Flowlet Switching for DCNIEEE Transactions on Cloud Computing10.1109/TCC.2024.338213212:2(580-593)Online publication date: Apr-2024
  • (2024)Extended Reality (XR) Codec Adaptation in 5G using Multi-Agent Reinforcement Learning with Attention Action Selection2024 IEEE 35th International Symposium on Personal, Indoor and Mobile Radio Communications (PIMRC)10.1109/PIMRC59610.2024.10817188(1-6)Online publication date: 2-Sep-2024
  • (2024)Discovery of False Data Injection Schemes on Frequency Controllers with Reinforcement Learning2024 IEEE Power & Energy Society General Meeting (PESGM)10.1109/PESGM51994.2024.10688906(1-5)Online publication date: 21-Jul-2024
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media