Abstract
Dynamic capacity management (or dynamic provisioning) is the process of dynamically changing the capacity allocation (reservation) of a virtual path (or a pseudo-wire) established between two network end points. This process is based on certain criteria including instantaneous traffic load for the pseudo-wire, network utilization, hour of day, or day of week. Frequent adjustment of the capacity yields a scalability issue in the form of a significant amount of message distribution and processing (i.e., signaling) in the network elements involved in the capacity update process. We therefore use the term “signaling rate” for the number of capacity updates per unit time. On the other hand, if the capacity is adjusted once and for the highest loaded traffic conditions, a significant amount of bandwidth may be wasted depending on the actual traffic load. There is then a need for dynamic capacity management that takes into account the tradeoff between signaling scalability and bandwidth efficiency. In this paper, we introduce a Markov decision framework for an optimal capacity management scheme. Moreover, for problems with large sizes and for which the desired signaling rate is imposed as a constraint, we provide suboptimal schemes using reinforcement learning. Our numerical results demonstrate that the reinforcement learning schemes that we propose provide significantly better bandwidth efficiencies than the static allocation policy without violating the signaling rate requirements of the underlying network.
This work is supported by The Scientific and Technical Research Council of Turkey (TUBITAK) under grant EEEAG-101E048
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
“ATM User Network Interface (UNI)”, ATM Forum Specification version 4.0, AFUNI-4.0, July 1996.
F. Baker, C. Iturralde, F. Le Faucheur, B. Davie, “Aggregation of RSVP for IPv4 and IPv6 Reservations”, RFC 3175, September 2001.
D. P. Bertsekas and J. N. Tsitsiklis, “Neuro-Dynamic Programming” Athena Scientific, Belmont, MA, 1996.
S. Blake, D. Black, M. Carlson, E. Davies, Z. Wang, W. Weiss, “An Architecture for Differentiated Services”, RFC 2475, 1998.
D. Clark, S. Shenker, L. Zhang, “Supporting Real-time Applications in an Integrated Services Packet Network: Architecture and Mechanism”, in Proc. SIGCOMM’92, September 1992.
B. Davie, Y. Rekhter, “MPLS: Technology and Applications”, Morgan Kaufmann Publishers, 2000.
A. Gosavi, “A Convergent Reinforcement Learning Algorithm for Solving Markov and Semi-Markov Decision Problems Under Long-Run Average Cost”, Accepted in the European Journal of Operational Research, 2001.
B. Groskinsky, D. Medhi, D. Tipper, “An Investigation of Adaptive Capacity Control Schemes in a Dynamic Traffic Environment”, IEICE Trans. Commun., Vol. E00-A, No, 13, 2001.
J. Heinanen, R. Guerin, “A Two Rate Three Color Marker”, RFC 2698, 1999.
A. Jalali, M. Ferguson. “Computationally efficient adaptive control algorithms for Markov chains”, In Proceedings of the 28th. IEEE Conference on Decision and Control, pages 1283–1288, 1989.
S. Mahadevan, “Average Reward Reinforcement Learning: Foundations, Algorithms and Empirical Results”, Machine Learning, 22, 159–196, 1996.
K. Nichols, S. Blake, F. Baker, D. Black, “Definition of the Differentiated Services Field (DS field) in the IPv4 and IPv6 Headers”, RFC 2474, December 1998.
S. Shenker, C. Partridge, R. Guerin, “Specification of Guaranteed Quality of Service”, RFC 2212, 1997.
S. Shiodam, H. Saito, H. Yokoi, “Sizing and Provisioning for Physical and Virtual Path Networks Using Self-sizing Capability”, IEICE Trans. Commun., Vol. E80-B, No. 2, February 1997.
R. S. Sutton and A. G. Barto, “Reinforcement Learning: An Introduction”, MIT Press, 1998.
H. C. Tijms, “Stochastic Models: An Algorithmic Approach”, John Wiley and Sons Ltd., 1994.
J. Wroclawski, “Specification of the Controlled-Load Network Element Service”, RFC 2211, 1997.
J. Wroclawski, “The Use of RSVP with IETF Integrated Services”, RFC 2210, 1997.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Akar, N., Sahin, C. (2003). Reinforcement Learning as a Means of Dynamic Aggregate QoS Provisioning. In: Burakowski, W., Bęben, A., Koch, B. (eds) Architectures for Quality of Service in the Internet. Art-QoS 2003. Lecture Notes in Computer Science, vol 2698. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45020-3_8
Download citation
DOI: https://doi.org/10.1007/3-540-45020-3_8
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40444-6
Online ISBN: 978-3-540-45020-7
eBook Packages: Springer Book Archive