Loading [a11y]/accessibility-menu.js
DMA cache: Using on-chip storage to architecturally separate I/O data from CPU data for improving I/O performance | IEEE Conference Publication | IEEE Xplore

DMA cache: Using on-chip storage to architecturally separate I/O data from CPU data for improving I/O performance


Abstract:

As technology advances both in increasing bandwidth and in reducing latency for I/O buses and devices, moving I/O data in/out memory has become critical. In this paper, w...Show More

Abstract:

As technology advances both in increasing bandwidth and in reducing latency for I/O buses and devices, moving I/O data in/out memory has become critical. In this paper, we have observed the different characteristics of I/O and CPU memory reference behavior, and found the potential benefits of separating I/O data from CPU data. We propose a DMA cache technique to store I/O data in dedicated on-chip storage and present two DMA cache designs. The first design, Decoupled DMA Cache (DDC), adopts additional on-chip storage as the DMA cache to buffer I/O data. The second design, Partition-Based DMA Cache (PBDC), does not require additional on-chip storage, but can dynamically use some ways of the processor's last level cache (LLC) as the DMA cache. We have implemented and evaluated the two DMA cache designs by using an FPGA-based emulation platform and the memory reference traces of real-world applications. Experimental results show that, compared with the existing snooping-cache scheme, DDC can reduce memory access latency (in bus cycles) by 34.8% on average (up to 58.4%), while PBDC can achieve about 80% of DDC's performance improvements despite no additional on-chip storage.
Date of Conference: 09-14 January 2010
Date Added to IEEE Xplore: 01 April 2010
ISBN Information:

ISSN Information:

Conference Location: Bangalore, India

References

References is not available for this document.