Skip to main content

Research Characterization on I/O Improvements of Storage Environments

  • Conference paper
  • First Online:
Advances on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC 2019)

Abstract

Nowadays, it has being verified some interesting improvements in I/O architectures. This is an essential point to complex and data intensive scalable applications. In the scientific and industrial fields, the storage component is a key element, because usually those applications employ a huge amount of data. Therefore, the performance of these applications commonly depends on some factors related to time spent in execution of the I/O operations. In this paper we present a research characterization on I/O improvements related to the storage targeting high-performance computing (HPC) and data-intensive scalable computing (DISC) applications. We also evaluated some of these improvements in order to justify their concerns with the I/O layer. Our experiments were performed in the Grid’5000, an interesting testbed distributed environment, suitable for better understanding challenges related to HPC and DISC applications. Results on synthetic I/O benchmarks, demonstrate how to improve the performance of the latency parameter for I/O operations.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Axboe, J.: Linux block IO–present and future. In: Ottawa Linux Symposium, pp. 51–61 (2004)

    Google Scholar 

  2. Bhattacharjee, B., Ross, K.A., Lang, C., Mihaila, G.A., Banikazemi, M.: Enhancing recovery using an SSD buffer pool extension. In: Proceedings of the Seventh International Workshop on Data Management on New Hardware, pp. 10–16. ACM (2011)

    Google Scholar 

  3. Chang, C., Greenwald, M., Riley, K., et al.: Fusion energy sciences exascale requirements review. an office of science review sponsored jointly by advanced scientific computing research and fusion energy sciences. In: USDOE Office of Science (SC) (2017)

    Google Scholar 

  4. Chang, L., Huang, S., Chou, K.: Relieving self-healing SSDs of heal storms. In: 10th ACM International Systems and Storage Conference, p. 5. ACM (2017)

    Google Scholar 

  5. Chen, R., Shen, Z., Ma, C., Shao, Z., Guan, Y.: NVMRA: utilizing NVM to improve the random write operations for NAND-flash-based mobile devices. Softw. Pract. Exp. 46, 1263–1284 (2016)

    Article  Google Scholar 

  6. Dorier, M., Antoniu, G., Cappello, F., Snir, M., Orf, L.: Damaris: how to efficiently leverage multicore parallelism to achieve scalable, jitter-free I/O. In: IEEE International Conference on Cluster Computing, pp. 155-163. IEEE (2012)

    Google Scholar 

  7. Dorier, M., Antoniu, G., Ross, R., Kimpe, D., Ibrahim, S.: CALCioM: mitigating i/o interference in HPC systems through cross-application coordination. In: IEEE 28th International Parallel and Distributed Processing Symposium, pp. 155–164. IEEE (2014)

    Google Scholar 

  8. Du, C., Wu, C., Li, J., Guo, M., He, X.: BPS: a balanced partial stripe write scheme to improve the write performance of raid-6 In: IEEE International Conference on Cluster Computing, pp. 204–213. IEEE (2015)

    Google Scholar 

  9. Gorton, I., Klein, J.: Distribution, data, deployment: software architecture convergence in big data systems. IEEE Softw. 32, 78–85 (2015)

    Article  Google Scholar 

  10. Huo, Z., Huo, X., et al.: A metadata cooperative caching architecture based on SSD and DRAM for file systems. In: International Conference on Algorithms and Architectures for Parallel Processing, pp. 31–51. Springer (2015)

    Google Scholar 

  11. Inacio, E.C. and Dantas, M.A.R.: IORE: a flexible and distributed i/o performance evaluation tool for hyperscale storage systems. In: Symposium on Computers and Communications (ISCC), pp. 01026–01031. IEEE (2018)

    Google Scholar 

  12. Kannan, S., Gavrilovska, A., Schwan, K., Milojicic, D., Talwar, V.: Using active NVRAM for I/O staging. In: Proceedings of the 2nd International Workshop on Petascal Data Analytics: Challenges and Opportunities, pp. 15–22. ACM (2011)

    Google Scholar 

  13. Kim, J., Ahn, S., La, K., Chang, W.: Improving I/O performance of NVMe SSD on virtual machines. In: Proceedings of the 31st Annual ACM Symposium on Applied Computing, pp. 1852–1857. ACM (2016)

    Google Scholar 

  14. Kuo, C., Shah, A., Nomura, A., Matsuoka, S., Wolf, F.: How file access patterns influence interference among cluster applications. In: International Conference on Cluster Computing (CLUSTER), pp. 185–193. IEEE (2014)

    Google Scholar 

  15. Lucas, R., Ang, J., Bergman k., et al.: Top ten exascale research challenges. DOE ASCAC subcommittee report, 1–86 (2014)

    Google Scholar 

  16. Mackey, G., Agun, M., Heinrich, M., Ryan, R., Yu, J.: Igloos make the cold bearable: a novel HDD technology for cold storage. In: 20th International Conference on HPC and Communications; 16th International Conference on Smart City; 4th International Conference on Data Science and Systems (HPCC/SmartCity/DSS), pp. 99–108. IEEE (2018)

    Google Scholar 

  17. Min, J., Ahn, S., La, K., Chang, W., Kim, J.: Cgroup++: enhancing I/O resource management of Linux Cgroup on NUMA systems with NVMe SSDs In: Proceedings of the Posters and Demos Session of the 16th International Middleware Conference, p. 7. ACM (2015)

    Google Scholar 

  18. Nakashima, K., Kon, J., Yamaguchi, S.: I/O performance improvement of secure big data analyses with application support on SSD cache. In: Proceedings of the 12th International Conference on Ubiquitous Information Management and Communication, p. 90. ACM (2018)

    Google Scholar 

  19. Ou, Y., Wu, X., Xiao, N., Liu, F., Chen, W.: NIS: a new index scheme for flash file system. In: 29th Symposium on Mass Storage Systems and Technologies (MSST), pp. 44–51. IEEE (2015)

    Google Scholar 

  20. Ouyang, X., Marcarelli, S., Panda, D.K.: Enhancing checkpoint performance with staging IO and SSD. In: International Workshop on Storage Network Architecture and Parallel I/Os, pp. 13–20. IEEE (2010)

    Google Scholar 

  21. Ramasamy, A.S., Karantharaj, P.: RFFE: a buffer cache management algorithm for flash-memory-based SSD to improve write performance. Can. J. Electr. Comput. Eng. 38, 219–231 (2015)

    Article  Google Scholar 

  22. Randal E. B.: Data-intensive supercomputing: intensive supercomputing: the case for DISC the case for DISC. Technical report: CMU-CS-07-128 (2019)

    Google Scholar 

  23. Shen, K., Park, S.: FlashFQ: a fair queueing I/O scheduler for flash-based SSDs. In: Presented as part of the 2013 USENIX Annual Technical Conference USENIX (ATC 2013), pp. 67–78. ACM (2013)

    Google Scholar 

  24. Shi, X., Li, M., Liu, W., Jin, H., Yu, C., Chen, Y.: SSDUP: a traffic-aware SSD burst buffer for HPC systems. In: Proceedings of the International Conference on Supercomputing, p. 27. ACM (2017)

    Google Scholar 

  25. Stratikopoulos, A., Kotselidis, C., Goodacre, J., Luján, M.: FastPath: towards wire-speed NVMe SSDs. In: 28th International Conference on Field Programmable Logic and Applications (FPL), pp. 170–1707. IEEE (2018)

    Google Scholar 

  26. Wan, L., Wolf, M., Wang, F., Choi, J.Y., Ostrouchov, G., Klasky, S.: Comprehensive measurement and analysis of the user-perceived I/O performance in a production leadership-class storage system. In: International Conference on Distributed Computing Systems (ICDCS), pp. 1022–1031. IEEE (2017)

    Google Scholar 

  27. Wu, C.H., et al.: A priority-based data placement method for databases using solid-state drives. In: Proceedings of the 2018 Conference on Research in Adaptive and Convergent Systems, pp. 175–182. ACM (2018)

    Google Scholar 

  28. Yung, C., Liu, X., Cheng, X.,: Content look-aside buffer for redundancy-free virtual disk I/O and caching. In: International Conference on Virtual Execution Environments, pp. 214–227. ACM (2017)

    Google Scholar 

  29. Yang, J., Pei S., Yang, Q.: WARCIP: write amplification reduction by clustering I/O pages. In: 12th ACM International Conference on Systems and Storage, pp. 155–166. ACM (2019)

    Google Scholar 

  30. Yildiz, O., Dorier, M., Ibrahim, S., Ross, R., Antoniu, G.: On the root causes of cross-application I/O interference in HPC storage systems. In: International Parallel and Distributed Processing Symposium (IPDPS), pp. 750–759. IEEE (2016)

    Google Scholar 

  31. Zhou, J., Chen, Y., Wang, W.: Atributed consistent hashing for heterogeneous storage systems. In: PACT, pp. 23–1. ACM (2018)

    Google Scholar 

Download references

Acknowledgment

Experiments presented in this paper were carried out using the Grid’5000 experimental testbed, being developed under the INRIA ALADDIN development action with support from CNRS, RENATER and several Universities as well as other funding bodies (see https://www.grid5000.fr). We also would like to thank the Federal University of Juiz de Fora (UFJF), CNPq, CAPES, FAPEMIG, PTI-LASSE and INESC P&D Brazil in SIGOM project that support in part this study.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Laércio Pioli .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Pioli, L., de Andrade Menezes, V.S., Dantas, M.A.R. (2020). Research Characterization on I/O Improvements of Storage Environments. In: Barolli, L., Hellinckx, P., Natwichai, J. (eds) Advances on P2P, Parallel, Grid, Cloud and Internet Computing. 3PGCIC 2019. Lecture Notes in Networks and Systems, vol 96. Springer, Cham. https://doi.org/10.1007/978-3-030-33509-0_26

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-33509-0_26

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-33508-3

  • Online ISBN: 978-3-030-33509-0

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics