skip to main content
10.1145/3575757.3593658acmotherconferencesArticle/Chapter ViewAbstractPublication PagesrtnsConference Proceedingsconference-collections
research-article

Reducing Overall Path Latency in Automotive Logical Execution Time Scheduling via Reinforcement Learning

Published:07 June 2023Publication History

ABSTRACT

The Logical Execution Time paradigm is a promising approach for achieving time-deterministic communication on multi-core CPUs. Task scheduling under this paradigm is a variant of the Multi-Row Facility Layout Problem, which is known to be NP-hard. In this paper, we propose using reinforcement learning to reduce the overall path latency among all scheduled runnables while adhering to other constraints, such as schedulability, load balance, and data contention control. The neural networks, also known as agents, are trained on a real-world automotive powertrain project. We compare two schedules generated by the agents to the current one and one produced by a genetic algorithm. The agent trained with the Proximal Policy Optimization algorithm demonstrated the best performance. Additionally, we investigate the generalization ability of the agents against software updates, and the results show that our agents are well-generalized.

References

  1. Kunal Agrawal, Sanjoy Baruah, Zhishan Guo, Jing Li, and Sudharsan Vaidhun. 2020. Hard-real-time routing in probabilistic graphs to minimize expected delay. In 2020 IEEE Real-Time Systems Symposium (RTSS). IEEE, 63–75.Google ScholarGoogle ScholarCross RefCross Ref
  2. Thomas Bäck, David B Fogel, and Zbigniew Michalewicz. 2018. Evolutionary computation 1: Basic algorithms and operators. CRC press.Google ScholarGoogle ScholarCross RefCross Ref
  3. Matthias Becker, Dakshina Dasari, Saad Mubeen, Moris Behnam, and Thomas Nolte. 2017. End-to-end timing analysis of cause-effect chains in automotive embedded systems. Journal of Systems Architecture 80 (2017), 104–113.Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Alessandro Biondi and Marco Di Natale. 2018. Achieving predictable multicore execution of automotive applications using the LET paradigm. In 2018 IEEE Real-Time and Embedded Technology and Applications Symposium (RTAS). IEEE, 240–250.Google ScholarGoogle ScholarCross RefCross Ref
  5. Zitong Bo, Ying Qiao, Chang Leng, Hongan Wang, Chaoping Guo, and Shaohui Zhang. 2021. Developing real-time scheduling policy by deep reinforcement learning. In 2021 IEEE 27th Real-Time and Embedded Technology and Applications Symposium (RTAS). IEEE, 131–142.Google ScholarGoogle ScholarCross RefCross Ref
  6. Greg Brockman, Vicki Cheung, Ludwig Pettersson, Jonas Schneider, John Schulman, Jie Tang, and Wojciech Zaremba. 2016. OpenAI Gym. arXiv:arXiv:1606.01540Google ScholarGoogle Scholar
  7. Maxwell Brown, Nicole Tellado, Sammy Kubesch, Denise Lainez, and Abdulmohsen Albelushi. 2023. Weighted Sum Method. https://optimization.cbe.cornell.edu/index.php?title=Weighted_sum_method.Google ScholarGoogle Scholar
  8. Tim Brys, Anna Harutyunyan, Peter Vrancx, Ann Nowé, and Matthew E Taylor. 2017. Multi-objectivization and ensembles of shapings in reinforcement learning. Neurocomputing 263 (2017), 48–59.Google ScholarGoogle ScholarCross RefCross Ref
  9. Sandeep Chinchali, Pan Hu, Tianshu Chu, Manu Sharma, Manu Bansal, Rakesh Misra, Marco Pavone, and Sachin Katti. 2018. Cellular network traffic scheduling with deep reinforcement learning. In Thirty-second AAAI conference on artificial intelligence.Google ScholarGoogle ScholarCross RefCross Ref
  10. Denis Claraz, Max J Friese, Hermann von Hasseln, and Ralph Mader. 2022. A dynamic Reference Architecture to achieve planned Determinism for Automotive Applications. In ERTS2022.Google ScholarGoogle Scholar
  11. Djork-Arné Clevert, Thomas Unterthiner, and Sepp Hochreiter. 2015. Fast and accurate deep network learning by exponential linear units (elus). arXiv preprint arXiv:1511.07289 (2015).Google ScholarGoogle Scholar
  12. Paul Emberson and Iain Bate. 2009. Stressing search with scenarios for flexible solutions to real-time task allocation problems. IEEE Transactions on Software Engineering 36, 5 (2009), 704–718.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Nico Feiertag, Kai Richter, Johan Nordlander, and Jan Jonsson. 2009. A compositional framework for end-to-end path delay calculation of automotive systems under different path semantics. In IEEE Real-Time Systems Symposium: 30/11/2009-03/12/2009. IEEE Communications Society.Google ScholarGoogle Scholar
  14. Anja Fischer, Frank Fischer, and Philipp Hungerländer. 2019. New exact approaches to row layout problems. Mathematical Programming Computation 11 (2019), 703–754.Google ScholarGoogle ScholarCross RefCross Ref
  15. Félix-Antoine Fortin, François-Michel De Rainville, Marc-André Gardner, Marc Parizeau, and Christian Gagné. 2012. DEAP: Evolutionary Algorithms Made Easy. Journal of Machine Learning Research 13 (jul 2012), 2171–2175.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Kai-Björn Gemlau, Johannes Schlatow, Mischa Möstl, and Rolf Ernst. 2017. Compositional Analysis of the WATERS Industrial Challenge 2017.Google ScholarGoogle Scholar
  17. Robert Glaubius, Terry Tidwell, Christopher Gill, and William D Smart. 2012. Real-time scheduling via reinforcement learning. arXiv preprint arXiv:1203.3481 (2012).Google ScholarGoogle Scholar
  18. Robert Glaubius, Terry Tidwell, William D Smart, and Christopher Gill. 2008. Scheduling design and verification for open soft real-time systems. In 2008 Real-Time Systems Symposium. IEEE, 505–514.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Aditya Gopalan and Shie Mannor. 2015. Thompson sampling for learning parameterized markov decision processes. In Conference on Learning Theory. PMLR, 861–898.Google ScholarGoogle Scholar
  20. Arne Hamann, Dakshina Dasari, Simon Kramer, Michael Pressler, and Falk Wurst. 2017. Communication centric design in complex automotive embedded systems. In 29th Euromicro Conference on Real-Time Systems (ECRTS 2017). Schloss Dagstuhl-Leibniz-Zentrum fuer Informatik.Google ScholarGoogle Scholar
  21. Julien Hennig, Hermann von Hasseln, Hassan Mohammad, Stefan Resmerita, Stefan Lukesch, and Andreas Naderlinger. 2016. Towards parallelizing legacy embedded control software using the LET programming paradigm. In 2016 IEEE Real-Time and Embedded Technology and Applications Symposium (RTAS). IEEE Computer Soc., 51.Google ScholarGoogle ScholarCross RefCross Ref
  22. Thomas A Henzinger, Benjamin Horowitz, and Christoph Meyer Kirsch. 2001. Giotto: A time-triggered language for embedded programming. In International Workshop on Embedded Software. Springer, 166–184.Google ScholarGoogle ScholarCross RefCross Ref
  23. Shengyi Huang and Santiago Ontañón. 2020. A closer look at invalid action masking in policy gradient algorithms. arXiv preprint arXiv:2006.14171 (2020).Google ScholarGoogle Scholar
  24. Shingo Igarashi, Tasuku Ishigooka, Tatsuya Horiguchi, Ryotaro Koike, and Takuya Azumi. 2020. Heuristic contention-free scheduling algorithm for multi-core processor using LET model. In 2020 IEEE/ACM 24th International Symposium on Distributed Simulation and Real Time Applications (DS-RT). IEEE, 1–10.Google ScholarGoogle ScholarCross RefCross Ref
  25. "Infineon Technologies AG" 2020. "Infineon AURIX TC3xx User Manual". "Infineon Technologies AG". "Rev. 1.6".Google ScholarGoogle Scholar
  26. Tobias Klaus, Matthias Becker, Wolfgang Schröder-Preikschat, and Peter Ulbrich. 2021. Constrained data-age with job-level dependencies: How to reconcile tight bounds and overheads. In 2021 IEEE 27th Real-Time and Embedded Technology and Applications Symposium (RTAS). IEEE, 66–79.Google ScholarGoogle ScholarCross RefCross Ref
  27. Alix Kordon and Ning Tang. 2020. Evaluation of the Age Latency of a Real-Time Communicating System using the LET paradigm. In ECRTS 2020, Vol. 165. Schloss Dagstuhl–Leibniz-Zentrum fuer Informatik.Google ScholarGoogle Scholar
  28. Ravi Kothari and Diptesh Ghosh. 2012. A competitive genetic algorithm for single row facility layout. (2012).Google ScholarGoogle Scholar
  29. Andrew Kusiak and Sunderesh S Heragu. 1987. The facility layout problem. European Journal of operational research 29, 3 (1987), 229–251.Google ScholarGoogle Scholar
  30. Ralph Mader. 2018. Implementation of Logical Execution Time in an AUTOSAR based Embedded Automotive Multi-core Application. In Dagstuhl Seminar 18092.Google ScholarGoogle Scholar
  31. Jorge Martinez, Ignacio Sañudo, and Marko Bertogna. 2018. Analytical characterization of end-to-end communication delays with logical execution time. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 37, 11 (2018), 2244–2254.Google ScholarGoogle ScholarCross RefCross Ref
  32. Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Alex Graves, Ioannis Antonoglou, Daan Wierstra, and Martin Riedmiller. 2013. Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602 (2013).Google ScholarGoogle Scholar
  33. Chris Nota and Philip S Thomas. 2019. Is the policy gradient a gradient?arXiv preprint arXiv:1906.07073 (2019).Google ScholarGoogle Scholar
  34. Hao Yi Ong, Kevin Chavez, and Augustus Hong. 2015. Distributed deep Q-learning. arXiv preprint arXiv:1508.04186 (2015).Google ScholarGoogle Scholar
  35. Paolo Pazzaglia, Alessandro Biondi, and Marco Di Natale. 2019. Optimizing the functional deployment on multicore platforms with logical execution time. In 2019 IEEE Real-Time Systems Symposium (RTSS). IEEE, 207–219.Google ScholarGoogle ScholarCross RefCross Ref
  36. Paolo Pazzaglia, Daniel Casini, Alessandro Biondi, and Marco Di Natale. 2021. Optimal memory allocation and scheduling for dma data transfers under the let paradigm. In 2021 58th ACM/IEEE Design Automation Conference (DAC). IEEE, 1171–1176.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. Antonin Raffin, Ashley Hill, Adam Gleave, Anssi Kanervisto, Maximilian Ernestus, and Noah Dormann. 2021. Stable-Baselines3: Reliable Reinforcement Learning Implementations. Journal of Machine Learning Research 22, 268 (2021), 1–8. http://jmlr.org/papers/v22/20-1364.htmlGoogle ScholarGoogle Scholar
  38. Stefan Resmerita, Andreas Naderlinger, Manuel Huber, Kenneth Butts, and Wolfgang Pree. 2015. Applying real-time programming to legacy embedded control software. In 2015 IEEE 18th International Symposium on Real-Time Distributed Computing. IEEE, 1–8.Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. Amir Sadrzadeh. 2012. A genetic algorithm with the heuristic procedure to solve the multi-line layout problem. Computers & Industrial Engineering 62, 4 (2012), 1055–1064.Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. John Schulman, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov. 2017. Proximal policy optimization algorithms. arXiv preprint arXiv:1707.06347 (2017).Google ScholarGoogle Scholar
  41. Richard S Sutton and Andrew G Barto. 2018. Reinforcement learning: An introduction. MIT press.Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. Hado Van Hasselt, Arthur Guez, and David Silver. 2016. Deep reinforcement learning with double q-learning. In Proceedings of the AAAI conference on artificial intelligence, Vol. 30.Google ScholarGoogle ScholarCross RefCross Ref
  43. Ziyu Wang, Tom Schaul, Matteo Hessel, Hado Hasselt, Marc Lanctot, and Nando Freitas. 2016. Dueling network architectures for deep reinforcement learning. In International conference on machine learning. PMLR, 1995–2003.Google ScholarGoogle Scholar
  44. Risheng Xu, Max J Friese, Hermann von Hasseln, and Dirk Nowotka. 2022. Data Access Time Estimation in Automotive LET Scheduling with Multi-core CPU. In 15th Junior Researcher Workshop on Real-Time Computing 2022. 11.Google ScholarGoogle Scholar

Index Terms

  1. Reducing Overall Path Latency in Automotive Logical Execution Time Scheduling via Reinforcement Learning

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      RTNS '23: Proceedings of the 31st International Conference on Real-Time Networks and Systems
      June 2023
      242 pages
      ISBN:9781450399838
      DOI:10.1145/3575757

      Copyright © 2023 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 7 June 2023

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

      Acceptance Rates

      Overall Acceptance Rate119of255submissions,47%
    • Article Metrics

      • Downloads (Last 12 months)78
      • Downloads (Last 6 weeks)5

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format