Abstract
The effectiveness of level one (L1) caches is of great importance to the processor performance. We have observed that programs exhibit varying demands in the L1 instruction cache (I-cache) and data cache (D-cache) during execution, and such demands are notably different across programs. We propose to co-allocate the cache ways between the I- and D-cache in responses to the program’s need on-the-fly. Resources are re-allocated based on the potential performance benefit. Using this scheme, a 32KB co-allocation L1 can gain 10% performance improvement on average, which is comparable to a 64KB traditional L1.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Alameldeen, A., Wood, D.: Adaptive cache compression for high-performance processors. In: ISCA, pp. 212–223 (2004)
Balasubramonian, R., Albonesi, D., Buyuktosunoglu, A., Dwarkadas, S.: Memory Hierarchy Reconfiguration for Energy and Performance in General-Purpose Processor Architectures. In: Micro, pp. 245–257 (2000)
Burger, D., Austin, T.: The SimpleScalar Tool Set, Version 2.0. Technical Report 1342, Univ. of Wisconsin-Madison, Comp. Sci. Dept. (1997)
Denning, P.: The working set model for program behavior. Communications of the ACM 11(5), 323–333 (1968)
Drach, N., Seznec, A.: Semi-unified caches. In: ICPP, pp. 25–28 (1993)
Karkhanis, T., Smith, J.: A first-order superscalar processor model. In: ISCA, pp. 338–349 (2004)
Kharbutli, M., Irwin, K., Solihin, Y., Lee, J.: Using prime numbers for cache indexing to eliminate conflict misses. In: HPCA, pp. 288–299 (2004)
Nesbit, K.J., Smith, J.E.: Data cache prefetching using a global history buffer. In: HPCA, pp. 96–105 (2004)
Ranganathan, P., Adve, S., Jouppi, N.: Reconfigurable caches and their application to media processing. In: ISCA, pp. 214–224 (2000)
Sherwood, T., Perelman, E., Hamerly, G., Calder, B.: Automatically Characterizing Large Scale Program Behavior. In: ASPLOS, pp. 45–57 (2002)
Shivakumar, P., Jouppi, N.P.: CACTI 3.0 An Integrated Cache Timing, Power, and Area Model. WRL Research Report (2001)
Suh, G.E., Devadas, S., Rudolph, L.: A new memory monitoring scheme for memory-aware scheduling and partitioning. In: HPCA, pp. 117–128 (2002)
Yang, J., Zhang, Y., Gupta, R.: Frequent value compression in data caches. In: MICRO, pp. 258–265 (2000)
Zhang, C., Vahid, F., Najjar, W.: A highly configurable cache architecture for embedded systems. In: ISCA, pp. 136–146 (2003)
Understanding the detailed architecture of the AMD’s 64 bit core, http://www.chip-architect.com/news/2003_09_21_Detailed_Architecture_of_AMDs_64bit_Core.html#3
Intel, Intel Pentium III Processor for the SC242 at 450 MHz to 1.0 GHz Datasheet, http://www.intel.com/design/pentiumiii/datashts/244452.htm
CPU Comparison, http://www.pantherproducts.co.uk/Articles/CPU/CPU%20Comparison.shtml
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Jin, L., Wu, W., Yang, J., Zhang, C., Zhang, Y. (2005). Dynamic Co-allocation of Level One Caches. In: Yang, L.T., Zhou, X., Zhao, W., Wu, Z., Zhu, Y., Lin, M. (eds) Embedded Software and Systems. ICESS 2005. Lecture Notes in Computer Science, vol 3820. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11599555_36
Download citation
DOI: https://doi.org/10.1007/11599555_36
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-30881-2
Online ISBN: 978-3-540-32297-9
eBook Packages: Computer ScienceComputer Science (R0)