ABSTRACT
Improving logistics efficiency is a challenging task in logistic systems, since planning the vehicle routes highly relies on the changing traffic conditions and diverse demand scenarios. However, most existing approaches either neglect the dynamic traffic environment or adopt manually designed rules, which fails to efficiently find a high-quality routing strategy. In this paper, we present a novel artificial intelligence (AI) based framework for logistic systems. This framework can simulate the spatio-temporal traffic conditions to form a dynamic environment in a data-driven manner. Under such a simulated environment, it adopts deep reinforcement learning techniques to intelligently generate the optimized routing strategy. Meanwhile, we also design an interactive frontend to visualize the simulated environment and routing strategies, which help operators evaluate the task performance. We will showcase the results of AI-based simulation and optimization in our demonstration.
- Said Dabia, Stefan Ropke, Tom Van Woensel, and Ton De Kok. 2013. Branch and price for the time-dependent vehicle routing problem with time windows. Transportation science, Vol. 47, 3 (2013), 380--396.Google Scholar
- Lu Duan, Yang Zhan, Haoyuan Hu, Yu Gong, Jiangwen Wei, Xiaodong Zhang, and Yinghui Xu. 2020. Efficiently solving the practical vehicle routing problem: A novel joint learning approach. In Proceedings of the 26th ACM SIGKDD international conference on knowledge discovery & data mining. 3054--3063.Google ScholarDigital Library
- Xiaomin Fang, Jizhou Huang, Fan Wang, Lingke Zeng, Haijin Liang, and Haifeng Wang. 2020. Constgat: Contextual spatial-temporal graph attention network for travel time estimation at baidu maps. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 2697--2705.Google ScholarDigital Library
- Tao-yang Fu and Wang-Chien Lee. 2019. Deepist: Deep image-based spatio-temporal network for travel time estimation. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management. 69--78.Google Scholar
- Maha Gmira, Michel Gendreau, Andrea Lodi, and Jean-Yves Potvin. 2021. Tabu search for the time-dependent vehicle routing problem with time windows on a road network. European Journal of Operational Research, Vol. 288, 1 (2021), 129--140.Google ScholarCross Ref
- Google. 2023. On-demand Rides and Deliveries Solution Documentation. https://developers.google.com/maps/documentation/transportation-logistics/mobility.Google Scholar
- Hitachi. 2019. Hitachi Digital Solution for Logistics. https://www.hitachi.co.jp/New/cnews/month/2019/02/0228.html.Google Scholar
- F Hooshmand and SA MirHassani. 2019. Time dependent green VRP with alternative fuel powered vehicles. Energy Systems, Vol. 10, 3 (2019), 721--756.Google ScholarCross Ref
- Rabie Jaballah, Marjolein Veenstra, Leandro C Coelho, and Jacques Renaud. 2021. The time-dependent shortest path and vehicle routing problem. INFOR: Information Systems and Operational Research, Vol. 59, 4 (2021), 592--622.Google ScholarCross Ref
- Guangyin Jin, Huan Yan, Fuxian Li, Yong Li, and Jincai Huang. 2023. Dual Graph Convolution Architecture Search for Travel Time Estimation. ACM Transactions on Intelligent Systems and Technology, Vol. 14, 4 (2023), 1--23.Google ScholarDigital Library
- Minsu Kim, Jinkyoo Park, et al. 2021. Learning collaborative policies to solve NP-hard routing problems. Advances in Neural Information Processing Systems, Vol. 34 (2021), 10418--10430.Google Scholar
- Wouter Kool, Herke Van Hoof, and Max Welling. 2018. Attention, learn to solve routing problems! arXiv preprint arXiv:1803.08475 (2018).Google Scholar
- Yeong-Dae Kwon, Jinho Choo, Byoungjip Kim, Iljoo Yoon, Youngjune Gwon, and Seungjai Min. 2020. Pomo: Policy optimization with multiple optima for reinforcement learning. Advances in Neural Information Processing Systems, Vol. 33 (2020), 21188--21198.Google Scholar
- Yeong-Dae Kwon, Jinho Choo, Iljoo Yoon, Minah Park, Duwon Park, and Youngjune Gwon. 2021. Matrix encoding networks for neural combinatorial optimization. Advances in Neural Information Processing Systems, Vol. 34 (2021), 5138--5149.Google Scholar
- Henry CW Lau, TM Chan, Wan Ting Tsui, and GTS Ho. 2009. Cost optimization of the supply chain network using genetic algorithms-Withdrawn. IEEE Transactions on Knowledge and Data Engineering (2009).Google Scholar
- Jingwen Li, Liang Xin, Zhiguang Cao, Andrew Lim, Wen Song, and Jie Zhang. 2021. Heterogeneous attentions for solving pickup and delivery problem via deep reinforcement learning. IEEE Transactions on Intelligent Transportation Systems, Vol. 23, 3 (2021), 2306--2315.Google ScholarCross Ref
- Wuman Luo, Haoyu Tan, Lei Chen, and Lionel M Ni. 2013. Finding time period-based most frequent path in big trajectory data. In Proceedings of the 2013 ACM SIGMOD international conference on management of data. 713--724.Google ScholarDigital Library
- Mahmood Rahmani, Erik Jenelius, and Haris N Koutsopoulos. 2013. Route travel time estimation using low-frequency floating car data. In 16th international ieee conference on intelligent transportation systems (itsc 2013). IEEE, 2292--2297.Google ScholarCross Ref
- Dong Wang, Junbo Zhang, Wei Cao, Jian Li, and Yu Zheng. 2018. When will you arrive? Estimating travel time based on deep neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 32.Google ScholarCross Ref
- Zonghan Wu, Shirui Pan, Guodong Long, Jing Jiang, and Chengqi Zhang. 2019. Graph Wavenet for Deep Spatial-Temporal Graph Modeling (IJCAI'19). 1907--1913.Google Scholar
- Huan Yan, Guangyin Jin, Deng Wang, Yue Liu, and Yong Li. 2022. Jointly Modeling Intersections and Road Segments for Travel Time Estimation via Dual Graph Convolutional Networks. In International Conference on Spatial Data and Intelligence. Springer, 19--34.Google ScholarDigital Library
- Zefang Zong, Meng Zheng, Yong Li, and Depeng Jin. 2022. MAPDP: Cooperative Multi-Agent Reinforcement Learning to Solve Pickup and Delivery Problems. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36. 9980--9988.Google ScholarCross Ref
Index Terms
- An AI-based Simulation and Optimization Framework for Logistic Systems
Recommendations
Multi-Agent Reinforcement Learning For Multi Vehicles One-commodity Vehicle Routing Problem
AbstractVehicle Routing Problem with Dynamic and Stochastic information (DS-VPR) is mathematical problem reflecting a huge set of real-world logistic problem. In this paper, we presented a novel method to handle DS-VRP. Our approach depends on multi-agent ...
A deep reinforcement learning-based approach for autonomous lane-changing velocity control in mixed flow of vehicle group level
AbstractAs an important driving behavior, lane-changing has a great impact on the safety and efficiency of traffic flow interacting with surrounding vehicles, especially in mixed traffic flows with autonomous vehicles and human-driven vehicles. This ...
Nash double Q-based multi-agent deep reinforcement learning for interactive merging strategy in mixed traffic
AbstractThe interaction between ramp and mainline vehicles plays a crucial role in merging areas, especially in the mixed-traffic environment. The driving behaviours of human drivers are uncertain and diverse, and the uncertainty makes it more complex ...
Comments