Performance tradeoffs in cache design
- 17 May 1988
- journal article
- Published by Association for Computing Machinery (ACM) in ACM SIGARCH Computer Architecture News
- Vol. 16 (2) , 290-298
- https://doi.org/10.1145/633625.52433
Abstract
Cache memories have become common across a wide range of computer implementations. To date, most analyses of cache performance have concentrated on time independent metrics, such as miss rate and traffic ratio. This paper presents a series of simulations that explore the interactions between various organizational decisions and program execution time. We investigate the tradeoffs between cache size and CPU/Cache cycle time, set associativity and cycle time, and between block size and main memory speed. The results indicate that neither cycle time nor cache size dominates the other across the entire design space. For common implementation technologies, performance is maximized when the size is increased to the 32KB to 128KB range with modest penalties to the cycle time. If set associativity impacts the cycle time by more than a few nanoseconds, it increases overall execution time. Since the block size and memory transfer rate combine to affect the cache miss penalty, the optimum block size is substantially smaller than that which minimizes the miss rate. Finally, the interdependence between optimal cache configuration and the main memory speed necessitates multi-level cache hierarchies for high performance uniprocessors.Keywords
This publication has 5 references indexed in Scilit:
- Line (Block) Size Choice for CPU Cache MemoriesIEEE Transactions on Computers, 1987
- Bibliography and reading on CPU cache memories and related topicsACM SIGARCH Computer Architecture News, 1986
- The IBM 3090 system: An overviewIBM Systems Journal, 1986
- Cache MemoriesACM Computing Surveys, 1982
- A Comparative Study of Set Associative Memory Mapping Algorithms and Their Use for Cache and Main MemoryIEEE Transactions on Software Engineering, 1978