Elsevier

Computer Networks

Volume 182, 9 December 2020, 107515
Computer Networks

Realtime mobile bandwidth prediction using LSTM neural network and Bayesian fusion

https://doi.org/10.1016/j.comnet.2020.107515Get rights and content

Abstract

With the increasing popularity of mobile Internet and the higher bandwidth requirement of mobile applications, user Quality of Experience (QoE) is particularly important. For applications requiring high bandwidth and low delay, such as video streaming, video conferencing, and online gaming, etc., if the future bandwidth can be estimated in advance, applications can leverage the estimation to adjust their data transmission strategies and significantly improve the user QoE. In this paper, we focus on accurate bandwidth prediction to improve user QoE. Specifically, We study realtime mobile bandwidth prediction in various mobile networking scenarios, such as subway and bus rides along different routes. The primary method used is Long Short Term Memory (LSTM) recurrent neural network. In individual scenarios, LSTM significantly improves the prediction accuracy of state-of-the-art prediction algorithms, such as Recursive Least Squares (RLS) by 12% in Root Mean Square Error (RMSE) and by 17% in Mean Average Error (MAE). We further developed Multi-Scale Entropy (MSE) to analyze the bandwidth patterns in different mobility scenarios and discuss its connection to the achieved accuracy. For practical applications, we developed Model Switching and Bayes Model Fusion to use pre-trained LSTM models for online realtime bandwidth prediction.

Introduction

We have witnessed the tremendous growth of mobile Internet traffic in the recent years. Users are increasingly spending more time on mobile apps and consuming more content on their mobile devices. The mobile traffic growth is expected to accelerate in the foreseeable future with the introduction of 5G wireless access and new media-rich applications, such as Virtual Reality and Augmented Reality. Providing better Quality of Experience (QoE) to mobile app users is of great significance. However, one main challenge for mobile app developers and content providers is the high volatility of mobile wireless connections. The physical channel quality of a mobile user is constantly affected by interference generated by other users, his/her own mobility, and signal blockages from static and dynamic blockers [1], [2]. The bandwidth available for a mobile session is ultimately determined by the adaptations cross the protocol stack, ranging from adaptive coding and modulation at PHY layer, cellular scheduling at data link layer, hand-overs between base stations, to TCP congestion control, etc. For many mobile apps involving user interactivity and/or multimedia content, e.g., gaming, conferencing, and video streaming, it is critical to accurately estimate the available bandwidth in realtime to deliver a high quality of user Quality-of-Experience (QoE). In the example of video streaming, many recent algorithms on Dynamic Adaptive Streaming over HTTP (DASH) optimize the video rate selection for upcoming video chunks based on the predicted TCP throughput in a future time window of several seconds [3], [4], [5]. If TCP throughput in the next few seconds can be accurately predicted, the selected video rate can maximize the delivery video quality and avoid video freeze. Interactive video conferencing has much tighter delay constraint than streaming. To avoid self-congestion, the available bandwidth on cellular links has to be accurately estimated in realtime, which is used to guide the realtime video coding and transmission strategies [6], [7]. Bandwidth overestimate will lead to long end-to-end video delay or freezing, and bandwidth underestimate will lead to unnecessarily poor perceptual video quality. Again, accurate realtime bandwidth prediction is crucial for delivering a good conferencing experience, especially in mobile networking scenarios.

In this paper, we study realtime mobile bandwidth prediction using Long Short Term Memory (LSTM) [8] recurrent neural network and Bayes fusion. Recent advances in Deep Learning have demonstrated that Recurrent Neural Networks (RNN) are powerful tools for sequence modeling and can learn temporal patterns in sequential data. RNNs have been widely used in Natural Language Processing (NLP), speech recognition, and time series processing [9], [10]. There are rich structures in realtime mobile network bandwidth evolution, due to user mobility patterns, wireless signal propagation laws, physical blockage models, and the well-defined behaviors of network protocols. This presents abundant opportunities for developing LSTM-based realtime mobile bandwidth estimation. The main idea is to offline train LSTM RNN models that capture the temporal patterns in various mobile networking scenarios. The trained LSTM RNN models will be used online to predict in realtime the network bandwidth within a short future time window. Specifically, we investigate the following research questions:

  • 1.

    How much prediction accuracy improvement can LSTM Deep Learning models bring over the conventional statistical prediction models?

  • 2.

    How predictable is realtime bandwidth at different prediction intervals under different mobility scenarios? Is the LSTM prediction accuracy dependent on specific mobility scenarios?

  • 3.

    Should one train a separate LSTM model for each mobility scenario, or train a universal LSTM model that can be used in different scenarios?

  • 4.

    How should one switch between different pre-trained models when user mobility pattern changes? Is it possible to fuse the predictions from multiple models to adapt to mobility pattern changes?

Towards answering these questions, we made the following contributions:

  • We conducted a mobile bandwidth measurement campaign to collect consecutive bandwidth traces in New York City. Our traces cover different transportation methods along different routes at different times of the day.1 The traces are described in detail in Section 4.

  • We developed LSTM models for realtime one-second ahead and multi-second ahead bandwidth predictions. Through extensive experiments on our own dataset and the HSDPA dataset [11], we demonstrated that LSTM significantly outperforms the existing realtime bandwidth prediction algorithms. Our LSTM models and their performance are presented in Sections 3 LSTM based realtime bandwidth prediction, 4 Data collection and performance evaluation.

  • We systematically evaluated the sensitivity of LSTM models to different mobility scenarios by comparing the accuracy of per-scenario, cross-scenario and universal predictions. Using Multi-Scale Entropy (MSE) analysis, we studied the connection between prediction accuracy and bandwidth regularity at different time scales. MSE also provides us with guidelines to explore cross-scenario bandwidth prediction. The analysis is presented in Section 5.

  • We designed model switching, which selects the best-performing LSTM model in recent history, and can converge to the model that fits the most to the current mobility scenario. We also designed model fusion, which generates predictions through Bayesian fusion of outputs of multiple models, and can make smooth transition from one mobility scenario to another. Model switching and fusion are presented in Section 6.

The rest of the paper is organized as the following. The related work on realtime bandwidth prediction is reviewed in Section 2. We formally define the realtime bandwidth prediction problem and introduce our LSTM based prediction models in Section 3. The performance of LSTM models is evaluated by a public dataset and our own dataset in Section 4. We conduct Multi-Scale Entropy analysis on our collected bandwidth traces and analyze the prediction accuracy in Section 5. In Section 6, we present Model Switching and Model Fusion to address mobility pattern changes. The paper is concluded with future work in Section 7.

Section snippets

Related work

Realtime bandwidth prediction has been a challenging problem for the networking community. A simple history-based TCP throughput estimation algorithm was proposed in [12]. Authors of [13] proposed to train a Support Vector Regress (SVR) model [14] to predict TCP throughput based on the measured packet loss rate, packet delay, and the size of the file to be transmitted. In the context of DASH video streaming, in [3], we adopted the prediction algorithm in [12] to guide realtime chunk rate

LSTM based realtime bandwidth prediction

In this section, we formulate the realtime bandwidth prediction problem, and introduce our LSTM prediction models. The key parameters and notations are listed in Table 1.

Datasets

It is critical to train and test LSTM models using large representative bandwidth datasets. We first used the HSDPA [11] dataset from the University of Oslo, Norway. It consists of cellular bandwidth traces collected on different transportation methods, including Train, Tram, Ferry, Car, Bus, and Metro. For each trace, it recorded the bandwidth and location every 1000 ms, and the duration for each trace ranges from 500 to 1000 s. However, we later found that the bandwidth traces are too short

Prediction accuracy analysis using multi-scale entropy

The predictability of a time series is determined by its complexity and the temporal correlation at different time scales. The traditional entropy measure can be used to quantify the randomness of a signal: the higher the entropy, the more random thus less predictable. However, the traditional entropy measure cannot model the signal complexity and temporal correlation at different time scales. Recently, Multi-Scale Entropy (MSE) [29] has been proposed to measure the complexity of physical and

Model selection and fusion

In the previous sections, we introduced and evaluated the feasibility of LSTM models on bandwidth prediction in mobile scenarios. In addition, the MSE analysis reveals the correlation between cross-scenario prediction accuracy and the similarity between mobility scenarios. For practical applications, a set of LSTM models will be trained offline for representative mobility scenarios. A subset of LSTM models will be chosen to run online to generate realtime prediction. The main questions are: (1)

Conclusion

In this paper, we studied realtime mobile bandwidth prediction using LSTM deep neural networks and Bayes fusion. We collected a rich set of bandwidth traces from different mobility scenarios. We developed LSTM recurrent neural network models to capture the temporal structures in mobile bandwidth traces for accurate prediction. For both next-second and multi-second ahead predictions, LSTM outperforms other state-of-the-art prediction algorithms, such as RLS, by 12% in Root Mean Square Error

CRediT authorship contribution statement

Lifan Mei: Conceptualization, Methodology, Software, Investigation, Resources, Data curation, Writing - original draft, Visualization, Project administration. Runchen Hu: Validation, Software. Houwei Cao: Conceptualization, Methodology. Yong Liu: Conceptualization, Methodology, Writing - review & editing, Visualization, Supervision, Project administration. Zifan Han: Funding acquisition, Conceptualization. Feng Li: Funding acquisition, Conceptualization. Jin Li: Funding acquisition,

Acknowledgments

We thank all the anonymous reviewers for the valuable comments. This work was supported in part by the Gift Fund from Huawei Technologies.

Declaration of competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Lifan Mei completed his Bachelor’s degree in Automation at Xi’An Jiaotong University (XJTU), Xi’An, China in 2015, and Master in Electrical and Computer Engineering at New York University (NYU) in 2017. He was a visiting student at UC Berkeley and National Chiao-Tung University (NCTU), Taiwan (RoC) in 2013. From 2017 to now, he is a Ph.D. Candidate at Electrical and Computer Engineering at New York University (NYU). His research focuses on Computer and Mobile Network, Bandwidth Prediction,

References (29)

  • ChoK. et al.

    Learning phrase representations using RNN encoder-decoder for statistical machine translation

    (2014)
  • HintonG. et al.

    Deep neural networks for acoustic modeling in speech recognition

    IEEE Signal Process. Mag.

    (2012)
  • http://home.ifi.uio.no/paalh/dataset/hsdpa-tcp-logs/,...
  • MirzaM. et al.

    A machine learning approach to TCP throughput prediction

  • Cited by (43)

    View all citing articles on Scopus

    Lifan Mei completed his Bachelor’s degree in Automation at Xi’An Jiaotong University (XJTU), Xi’An, China in 2015, and Master in Electrical and Computer Engineering at New York University (NYU) in 2017. He was a visiting student at UC Berkeley and National Chiao-Tung University (NCTU), Taiwan (RoC) in 2013. From 2017 to now, he is a Ph.D. Candidate at Electrical and Computer Engineering at New York University (NYU). His research focuses on Computer and Mobile Network, Bandwidth Prediction, Routing Optimization, Machine Learning etc.

    View full text