Skip to main content

“Hey CAI” - Conversational AI Enabled User Interface for HPC Tools

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13289))

Abstract

HPC system users depend on profiling and analysis tools to obtain insights into the performance of their applications and tweak them. The complexity of modern HPC systems have necessitated advances in the associated HPC tools making them equally complex with various advanced features and complex user interfaces. While these interfaces are extensive and detailed, they require a steep learning curve even for expert users making them harder to use for novice users. While users are intuitively able to express what they are looking for in words or text (e.g., show me the process transmitting maximum data), they find it hard to quickly adapt to, navigate, and use the interface of advanced HPC tools to obtain desired insights. In this paper, we explore the challenges associated with designing a conversational (speech/text) interface for HPC tools. We use state-of-the-art AI models for speech and text and adapt it for use in the HPC arena by retraining them on a new HPC dataset we create. We demonstrate that our proposed model, retrained with an HPC specific dataset, can deliver higher accuracy than the existing state-of-the-art pre-trained language models. We also create an interface to convert speech/text data to commands for HPC tools and show how users can utilize the proposed interface to gain insights quicker leading to better productivity.

To the best of our knowledge, this is the first effort aimed at designing a conversational interface for HPC tools using state-of-the-art AI techniques to enhance the productivity of novice and advanced users alike.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Change history

  • 29 May 2022

    In an older version of this paper, the name of the fourth author was missing. This has been corrected.

References

  1. Horovod: Distributed training framework for TensorFlow. https://github.com/uber/horovod

  2. MPIP: Lightweight, Scalable MPI Profiling. http://www.llnl.gov/CASC/mpip/

  3. MVAPICH: MPI over InfiniBand, Omni-Path, Ethernet/iWARP, and RoCE. http://mvapich.cse.ohio-state.edu/features/. Accessed 13 March 2022

  4. Prometheus exporter. https://github.com/prometheus/node_exporter

  5. The future of conversational AI (2021). https://www2.deloitte.com/us/en/insights/focus/signals-for-strategists/the-future-of-conversational-ai.html. Accessed 13 March 2022

  6. Nvidia Nsight Developer Tools (2022). https://developer.nvidia.com/tools-overview. Accessed 13 March 2022

  7. The impact of voice assistants (2022). https://www.pwc.com/us/en/services/consulting/library/consumer-intelligence-series/voice-assistants.html. Accessed 13 March 2022

  8. Kousha, P., et al.: Accelerated real-time network monitoring and profiling at scale using OSU INAM. In: Practice and Experience in Advanced Research Computing (PEARC 2020) (2020)

    Google Scholar 

  9. Agelastos, A., et al.: The Lightweight Distributed Metric Service: A Scalable Infrastructure for Continuous Monitoring of Large Scale Computing Systems and Applications, pp. 154–165. SC 2014, IEEE Press, Piscataway, NJ, USA (2014). https://doi.org/10.1109/SC.2014.18, http://dx.doi.org/10.1109/SC.2014.18

  10. Amodei, D., et al.: Deep Speech 2: End-to-End Speech Recognition in English and Mandarin. CoRR abs/1512.02595 (2015). http://arxiv.org/abs/1512.02595

  11. Barth, B., Evans, T., McCalpin, J.: Tacc stats. https://www.tacc.utexas.edu/research-development/tacc-projects/tacc-stats

  12. Baevski, A., Zhou, H., Mohamed, A., Auli, M.: Wav2vec 2.0: a framework for self-supervised learning of speech representations (2020). https://arxiv.org/abs/2006.11477

  13. Castellucci, G., Bellomaria, V., Favalli, A., Romagnoli, R.: Multi-lingual intent detection and slot filling in a joint bert-based model. arXiv preprint arXiv:1907.02884 (2019)

  14. Coucke, A., et al.: Snips voice platform: an embedded spoken language understanding system for private-by-design voice interfaces. CoRR abs/1805.10190 (2018). http://arxiv.org/abs/1805.10190

  15. Garofolo, J., et al.: TIMIT Acoustic-Phonetic Continuous Speech Corpus (1993). 11272.1/AB2/SWVENO, https://hdl.handle.net/11272.1/AB2/SWVENO

  16. Hemphill, C.T., Godfrey, J.J., Doddington, G.R.: The ATIS spoken language systems pilot corpus, pp. 96–101. HLT 1990, Association for Computational Linguistics, USA (1990). https://doi.org/10.3115/116580.116613

  17. Hosseini-Asl, E., McCann, B., Wu, C., Yavuz, S., Socher, R.: A simple language model for task-oriented dialogue. CoRR abs/2005.00796 (2020). https://arxiv.org/abs/2005.00796

  18. HPCToolkit: (2019). http://hpctoolkit.org/. Accessed 13 March 2022

  19. Kousha, P., et al.: INAM: Cross-Stack Profiling and Analysis of Communication in MPI-Based Applications. Association for Computing Machinery, New York, NY, USA (2021). https://doi.org/10.1145/3437359.3465582

  20. Kudo, T., Richardson, J.: Sentencepiece: a simple and language independent subword tokenizer and detokenizer for neural text processing. arXiv preprint arXiv:1808.06226 (2018)

  21. Lugosch, L., Ravanelli, M., Ignoto, P., Tomar, V.S., Bengio, Y.: Speech model pre-training for end-to-end spoken language understanding (2019)

    Google Scholar 

  22. Malony, A.D., Shende, S.: Performance technology for complex parallel and distributed systems. In: Kotsis, G., Kacsuk, P. (eds.) Proceedings of DAPSYS 2000, pp. 37–46 (2000)

    Google Scholar 

  23. OSU InfiniBand Network Analysis and Monitoring Tool. http://mvapich.cse.ohio-state.edu/tools/osu-inam/

  24. Palogiannidi, E., Gkinis, I., Mastrapas, G., Mizera, P., Stafylakis, T.: End-to-end architectures for ASR-free spoken language understanding. In: ICASSP 2020–2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 7974–7978 (2020). https://doi.org/10.1109/ICASSP40776.2020.9054314

  25. Panayotov, V., Chen, G., Povey, D., Khudanpur, S.: Librispeech: an ASR corpus based on public domain audio books. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5206–5210. IEEE (2015)

    Google Scholar 

  26. Paszke, A., et al.: PyTorch: An Imperative Style, High-Performance Deep Learning Library (2019)

    Google Scholar 

  27. Qin, L., Che, W., Li, Y., Wen, H., Liu, T.: A stack-propagation framework with token-level intent detection for spoken language understanding. arXiv preprint arXiv:1909.02188 (2019)

  28. Serdyuk, D., Wang, Y., Fuegen, C., Kumar, A., Liu, B., Bengio, Y.: Towards end-to-end spoken language understanding. CoRR abs/1802.08395 (2018). http://arxiv.org/abs/1802.08395

  29. Sergeev, A., Balso, M.D.: Horovod: fast and easy distributed deep learning in TensorFlow. CoRR abs/1802.05799 (2018). http://arxiv.org/abs/1802.05799

  30. Wang, C., Tang, Y., Ma, X., Wu, A., Okhonko, D., Pino, J.: Fairseq s2t: fast speech-to-text modeling with fairseq (2020). https://arxiv.org/abs/2010.05171

  31. Wen, T., et al.: A network-based end-to-end trainable task-oriented dialogue system. CoRR abs/1604.04562 (2016). http://arxiv.org/abs/1604.04562

  32. Wu, D., Ding, L., Lu, F., Xie, J.: Slotrefine: a fast non-autoregressive model for joint intent detection and slot filling. CoRR abs/2010.02693 (2020). https://arxiv.org/abs/2010.02693

Download references

Acknowledgement

This research is supported in part by NSF grants #1818253, #1854828, #1931537, #2007991, #2018627, #2112606, and XRAC grant #NCR-130002.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Pouya Kousha .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Kousha, P. et al. (2022). “Hey CAI” - Conversational AI Enabled User Interface for HPC Tools. In: Varbanescu, AL., Bhatele, A., Luszczek, P., Marc, B. (eds) High Performance Computing. ISC High Performance 2022. Lecture Notes in Computer Science, vol 13289. Springer, Cham. https://doi.org/10.1007/978-3-031-07312-0_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-07312-0_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-07311-3

  • Online ISBN: 978-3-031-07312-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics