skip to main content
10.1145/1007912.1007950acmconferencesArticle/Chapter ViewAbstractPublication PagesspaaConference Proceedingsconference-collections
Article

Online algorithms for prefetching and caching on parallel disks

Published:27 June 2004Publication History

ABSTRACT

Parallel disks provide a cost effective way of speeding up I/Os in applications that work with large amounts of data. The main challenge is to achieve as much parallelism as possible, using prefetching to avoid bottlenecks in disk access. Efficient algorithms have been developed for some particular patterns of accessing the disk blocks. In this paper, we consider general request sequences. When the request sequence consists of unique block requests, the problem is called prefetching and is a well-solved problem for arbitrary request sequences. When the reference sequence can have repeated references to the same block, we need to devise an effective caching policy as well. While optimum offline algorithms have been recently designed for the problem, in the online case, no effective algorithm was previously known. Our main contribution is a deterministic online algorithm threshold-LRU which achieves O((MD/L)2/3) competitive ratio and a randomized online algorithm threshold-MARK which achieves O(√(MD/L) log (MD/L)) competitive ratio for the caching/prefetching problem on the parallel disk model (PDM), where D is the number of disks, M is the size of fast memory buffer, and M+L is the amount of lookahead available in the request sequence. The best-known lower bound on the competitive ratio is Ω(≾MD/L) for lookahead LM in both models. We also show that if the deterministic online algorithm is allowed to have twice the memory of the offline then a tight competitive ratio of Θ(≾MD/L) can be achieved. This problem generalizes the well-known paging problem on a single disk to the parallel disk model.

References

  1. S. Albers. On the influence of lookahead in competitive paging algorithms. Algorithmica, 18(3):283--305, 1997.]]Google ScholarGoogle ScholarCross RefCross Ref
  2. S. Albers and M. Büttner. Integrated prefetching and caching in single and parallel disk systems. In SPAA, pages 109--117, 2003.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. S. Albers, N. Garg, and S. Leonardi. Minimizing stall time in single and parallel disk systems. In In Proc. of 30th Annual ACM Symp. on Theory of Computing (STOC 98), pages 454--462, 1998.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. S. Albers and C. Witt. Minimizing stall time in single and parallel disk systems using multicommodity network flows. In RANDOM-APPROX, 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. R. Barve, M. Kallahalla, P. J. Varman, and J. S. Vitter. Competitive parallel disk prefetching and buffer management. In In Proc. of Fifth Workshop on I/O in parallel and Distributed Systems, pages 47--56, Nov 1997.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. L. A. Belady. A study of replacement algorithms for virtual storage computers. IBM Systems Journal, 5:78--101, 1966.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. A. Borodin and R. El-Yaniv. Online computation and competitive analysis. Cambridge University Press, 1998.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. D. Breslauer. On competitive online paging with lookahead. TCS, 290(1-2):365--375, 1998.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. P. Cao, E. W. Felton, A. R. Karlin, and K. Li. A study of integrated prefetching and caching strategies. In In Proc. of the joint Intl. Conf. on measurement and modeling of computer systems, pages 188--197, May 1995.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. A. Fiat, R. Karp, M. Luby, L. McGoech, D. D. Sleator, and N. E. Young. Competitive paging algorithms. Journal of Algorithms, 12(4):685--699, Dec 1991.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. D. A. Hutchinson, P. Sanders, and J. S. Vitter. Duality between prefetching and queued writing with application to integrated caching and prefetching and to external sorting. In ESA, 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. M. Kallahalla and P. J. Varman. Optimal read-once parallel disk scheduling. In In Proc. of Sixth ACM Workshop on I/O in Parallel and Distributed Systems, pages 68--77, 1999.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. M. Kallahalla and P. J. Varman. Optimal prefetching and caching for parallel i/o systems. In SPAA, 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. A. R. Karlin, M. S. Manasse, L. Rudolph, and D. D. Sleator. Competitive snoopy caching. Algorithmica, 3(1):79--119, 1988.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. T. Kimbrel, P. Cao, E.W. Felten, A. R. Karlin, and K. Li. Integrated parallel prefetching and caching. In SIGMETRICS, 1996.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. T. Kimbrel and A. R. Karlin. Near optimal parallel prefetching and caching. In FOCS, pages 540--549, 1996.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  17. L. A. McGeoch and D. D. Sleator. A strongly competitive randomized paging algorithm. Algorithmica, 6:816--825, 1991.]]Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. D. D. Sleator and R. E. Tarjan. Amortized efficiency of the list update and paging rules. Communications of the ACM, 28:202--208, November 1985.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. J. S. Vitter. External memory algorithms and data structures: Dealing with massive data. ACM Computing surveys, 33(2):209--271, June 2001.]] Google ScholarGoogle ScholarDigital LibraryDigital Library
  20. N. Young. Competitive paging and dual-guided on-line weighted caching and matching algorithms. In Ph.D. thesis. Princeton University, 1991. CS-TR-348-91.]] Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Online algorithms for prefetching and caching on parallel disks

        Recommendations

        Comments

        Login options

        Check if you have access through your login credentials or your institution to get full access on this article.

        Sign in
        • Published in

          cover image ACM Conferences
          SPAA '04: Proceedings of the sixteenth annual ACM symposium on Parallelism in algorithms and architectures
          June 2004
          332 pages
          ISBN:1581138407
          DOI:10.1145/1007912

          Copyright © 2004 ACM

          Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

          Publisher

          Association for Computing Machinery

          New York, NY, United States

          Publication History

          • Published: 27 June 2004

          Permissions

          Request permissions about this article.

          Request Permissions

          Check for updates

          Qualifiers

          • Article

          Acceptance Rates

          Overall Acceptance Rate447of1,461submissions,31%

          Upcoming Conference

          SPAA '24

        PDF Format

        View or Download as a PDF file.

        PDF

        eReader

        View online with eReader.

        eReader