ABSTRACT
This paper proposes a modified two-way set associative cache for embedded systems to reduce the energy consumption. For this goal, the proposed cache, called SSA (selective-way-access skewed associative) cache, equips with a way-selecting mechanism controlled by skewing function and small table look-up, which also has the reconfigurable ability to be converted to one direct mapped cache on a specific application. The skewing mechanism including differentiated mapping function for each cache set, and specialized replacement policy enables the perfect speculation in way-selection and decreases conflict misses. Consequently, the proposed cache effectively achieves the energy reduction without any performance degradation. Additional delay of a small table look-up for the way selection can be hidden partly and multiplexer delay in critical path can be removed totally, such that overall cache access time becomes almost same as that of conventional set associative cache. The simulation result shows that the proposed cache structure reduces energy consumption up to 30~55% over conventional set associative cache and up to 25~30% over previous way-prediction caches. Furthermore, the software controlled reconfigurable architecture brings flexibility with the proposed cache to operate as direct mapped cache or way selecting cache based on given application adaptively.
- A. Seznec, "A case For Two-way Skewed-associative Caches," Proceedings of the 20th Annual International Symposium on Computer Architecture, May 1993, pp. 169--178. Google ScholarDigital Library
- F. Bodin and A. Seznec, "Skewed associativity improves program performance and enhances predictability," IEEE Transactions on Computers, Volume: 46 Issue: 5, May 1997, pp. 530--544. Google ScholarDigital Library
- B. Calder, D. Grunwald, and J. Emer, "Predictive sequential associative cache," Proceedings of Second International Symposium on High-Performance Computer Architecture, Feb. 1996, pp. 244--253. Google ScholarDigital Library
- D. H. Albonesi, "Selective cache ways: on-demand cache resource allocation," Proceedings of 32nd Annual Int ernational Symposium on Microarchitecture. MICRO-32, Nov. 1999, pp. 248--259. Google ScholarDigital Library
- K. Inoue, T. Ishihara, and K. Murakami, "Way-predicting set-associative cache for high performance and low energy consumption," Proceedings of International Symposium on Low Power Electronics and Design, Aug. 1999, pp. 273--275. Google ScholarDigital Library
- S. J. E. Wilton and N. P. Jouppi, "CACTI: an enhanced cache access and cycle time model," IEEE Journal of Solid-State Circuits, Volume: 31 Issue: 5, May 1996, pp. 677--688.Google ScholarCross Ref
- T. Wada, S. Rajan, and S. A. Przybylski, "An analytical access time model for on-chip cache memories," IEEE Journal of Solid-State Circuits, Volume: 27 Issue: 8, Aug. 1992, pp. 1147--1156.Google ScholarCross Ref
- M. D. Powell, A. Agarwall, T. N. Vijaykumar, B. Falsafi, and K. Roy, "Reducing set-associative cache energy via way-prediction and selective direct-mapping," Proceedings of 34th ACM/IEEE International Symposium on Microarchitecture, MICRO-34. Dec. 2001, pp. 54--65. Google ScholarDigital Library
- M. B. Kamble and K. Ghose, "Energy-efficiency of VLSI caches: a comparative study," Proceedings of Tenth International Conference on VLSI Design, Jan. 1997, pp. 261--267. Google ScholarDigital Library
- M. B. Kamble and K. Ghosse, "Analytical energy dissipation models for low power caches," Proceedings of International Symposium on Low Power Electronics and Design, Aug. 1997, pp. 143--148. Google ScholarDigital Library
- J. H. Lee, S. W. Jeong, S. D. Kim, and C. Weems, "An Intelligent Cache System with Hardware prefetching for High Performance," IEEE Transaction on Computers, Vol. 52, No. 5, May 2003, pp. 607--616. Google ScholarDigital Library
- J. H. Lee, S. D. Kim, and C. Weems, "Application Adaptive Intelligent Cache Memory System," ACM Transactions on Embedded Computing Systems, Vol. 1, Issue. 1, Dec. 2002, pp. 56--78. Google ScholarDigital Library
- N. P. Jouppi, "Improving direct-mapped cache performance by the addition of a small fully-associative cache and prefetch buffers," Proceedings of 17th Annual International Symposium on Computer Architecture, May 1990, pp. 364--373. Google ScholarDigital Library
- Chuanjun Zhang, F. Vahid, and W. Najjar, "A highly configurable cache architecture for embedded systems," Proceedings of 30th Annual International Symposium on Computer Architecture, June. 2003, pp. 136--146. Google ScholarDigital Library
- J. Edler and M. D. Hill, Dinero IV Trace-Driven Uniprocessor Cache Simulator. Univ. Wisconsin; http://www.cs.wisc.edu/~markhill/DineroIV.Google Scholar
- D. Burger, T. M. Austin, "The SimpleScalar tool set, version 2.0," Technical Report TR-97-1342, University of Wisconsin-Madison, 1997.Google Scholar
- http://www.specbench.org/osg/cpu2000Google Scholar
- C. Lee, M. Potkonjak, and W. H. Mangione-Smith, "MediaBench: a tool for evaluating and synthesizing multimedia and communications systems", Proceedings of Thirtieth Annual IEEE/ACM International Symposium on Microarchitecture, Dec. 1997, pp. 330--335. Google ScholarDigital Library
- J. L. Hennessy, D. A. Patterson "Computer architecture -A Quantitative approach" Google ScholarDigital Library
- S. Kim, N. Vijaykrishnan, M. J. Irwin and L. K. John "On load latency in low-power caches", Proceedings of International Symposium on Low Power Electronics and Design, Aug. 2003 pp. 258--261 Google ScholarDigital Library
Index Terms
- An energy efficient cache memory architecture for embedded systems
Recommendations
An efficient direct mapped instruction cache for application-specific embedded systems
CODES+ISSS '05: Proceedings of the 3rd IEEE/ACM/IFIP international conference on Hardware/software codesign and system synthesisCaches may consume half of a microprocessor's total power and cache misses incur accessing off-chip memory, which is both time consuming and energy costly. Therefore, minimizing cache power consumption and reducing cache misses are important to reduce ...
An energy-delay efficient 2-level data cache architecture for embedded system
ISLPED '09: Proceedings of the 2009 ACM/IEEE international symposium on Low power electronics and designWe propose a 2-level data cache architecture with a low energy-delay product tailored for the embedded systems. The L1 data cache is small and direct-mapped, and employs a write-through policy. In contrast, the L2 data cache is set-associative and ...
Application-adaptive intelligent cache memory system
This article presents the design of a simple hardware-controlled, high performance cache system. The design supports fast access time, optimal utilization of temporal and spatial localities adaptive to given applications, and a simple dynamic fetching ...
Comments