Skip to main content

Exploring High Bandwidth Pipelined Cache Architecture for Scaled Technology

  • Chapter
Embedded Software for SoC

Abstract

In this article we propose a design technique to pipeline cache memories for high bandwidth applications. With the scaling of technology cache access latencies are multiple clock cycles. The proposed pipelined cache architecture can be accessed every clock cycle and thereby, enhances bandwidth and overall processor performance. The proposed architecture utilizes the idea of banking to reduce bit-line and word-line delay, making word-line to sense amplifier delay to fit into a single clock cycle. Experimental results show that optimal banking allows the cache to be split into multiple stages whose delays are equal to clock cycle time. The proposed design is fully scalable and can be applied to future technology generations. Power, delay and area estimates show that on average, the proposed pipelined cache improves MOPS (millions of operations per unit time per unit area per unit energy) by 40–50% compared to current cache architectures.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 249.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. K. Naogami, T. Sakurai et. al. “A 9-ns Hit-Delay 32-kbyte Cache Macro for High Speed RISC.” IEEE Journal of Solid State Circuits, Vol. 25, No. 1. February 1990.

    Google Scholar 

  2. T. Wada and S. Rajan. “An Analytical Access Time Model for On-Chip cache Memories.” IEEE Journal of Solid State Circuits, Vol. 27, No. 8, pp. 1147–1156, August 1992.

    Article  Google Scholar 

  3. J. L. Hennessy and D. A. Patterson. Computer Architecture A Quantitative Approach. Morgan KaufMann, 2nd Edition.

    Google Scholar 

  4. S. J. E. Wilson and N. P. Jouppi. “An Enhanced Access and Cycle Time Model for On-Chip Caches.” Technical Report 93/5, Digital Equipment Corporation, Western Research Laboratory, July 1994.

    Google Scholar 

  5. J. M. Rabaey. Digital Integrated Circuit. Prentice Hall, 1996.

    Google Scholar 

  6. D. M. Tullsen, S. J. Eggers, and H. M. Levy. “Simultaneous Multithreading: Maximizing On-Chip Parallelism.” In Proceedings of the 22th Annual International Symposium on Computer Architecture, pp. 392–403, June 1995.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2003 Kluwer Academic Publishers

About this chapter

Cite this chapter

Agarwal, A., Roy, K., Vijaykumar, T.N. (2003). Exploring High Bandwidth Pipelined Cache Architecture for Scaled Technology. In: Jerraya, A.A., Yoo, S., Verkest, D., Wehn, N. (eds) Embedded Software for SoC. Springer, Boston, MA. https://doi.org/10.1007/0-306-48709-8_26

Download citation

  • DOI: https://doi.org/10.1007/0-306-48709-8_26

  • Publisher Name: Springer, Boston, MA

  • Print ISBN: 978-1-4020-7528-5

  • Online ISBN: 978-0-306-48709-5

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics