By: Aidahani Binti Ahmad

Slides:



Advertisements
Similar presentations
1 Improving Direct-Mapped Cache Performance by the Addition of a Small Fully-Associative Cache and Prefetch Buffers By Sreemukha Kandlakunta Phani Shashank.
Advertisements

CSC 4250 Computer Architectures December 8, 2006 Chapter 5. Memory Hierarchy.
CS2100 Computer Organisation Cache II (AY2014/2015) Semester 2.
July 2005Computer Architecture, Memory System DesignSlide 1 Part V Memory System Design.
1 Lecture 20: Cache Hierarchies, Virtual Memory Today’s topics:  Cache hierarchies  Virtual memory Reminder:  Assignment 8 will be posted soon (due.
The Lord of the Cache Project 3. Caches Three common cache designs: Direct-Mapped store in exactly one cache line Fully Associative store in any cache.
Spring 2003CSE P5481 Introduction Why memory subsystem design is important CPU speeds increase 55% per year DRAM speeds increase 3% per year rate of increase.
Overview of Cache and Virtual MemorySlide 1 The Need for a Cache (edited from notes with Behrooz Parhami’s Computer Architecture textbook) Cache memories.
Computer ArchitectureFall 2007 © November 14th, 2007 Majd F. Sakr CS-447– Computer Architecture.
1 Lecture 12: Cache Innovations Today: cache access basics and innovations (Sections )
1 COMP 206: Computer Architecture and Implementation Montek Singh Mon, Oct 31, 2005 Topic: Memory Hierarchy Design (HP3 Ch. 5) (Caches, Main Memory and.
EENG449b/Savvides Lec /13/04 April 13, 2004 Prof. Andreas Savvides Spring EENG 449bG/CPSC 439bG Computer.
Csci4203/ece43631 Review Quiz. 1)It is less expensive 2)It is usually faster 3)Its average CPI is smaller 4)It allows a faster clock rate 5)It has a simpler.
1  Caches load multiple bytes per block to take advantage of spatial locality  If cache block size = 2 n bytes, conceptually split memory into 2 n -byte.
Computer ArchitectureFall 2008 © November 3 rd, 2008 Nael Abu-Ghazaleh CS-447– Computer.
Cache intro CSE 471 Autumn 011 Principle of Locality: Memory Hierarchies Text and data are not accessed randomly Temporal locality –Recently accessed items.
Prof. John Nestor ECE Department Lafayette College Easton, Pennsylvania ECE Computer Organization Memory Hierarchy 2.
Reducing Cache Misses 5.1 Introduction 5.2 The ABCs of Caches 5.3 Reducing Cache Misses 5.4 Reducing Cache Miss Penalty 5.5 Reducing Hit Time 5.6 Main.
Cache Memories Effectiveness of cache is based on a property of computer programs called locality of reference Most of programs time is spent in loops.
Cache memory October 16, 2007 By: Tatsiana Gomova.
 Higher associativity means more complex hardware  But a highly-associative cache will also exhibit a lower miss rate —Each set has more blocks, so there’s.
Memory/Storage Architecture Lab Computer Architecture Memory Hierarchy.
Lecture Objectives: 1)Define set associative cache and fully associative cache. 2)Compare and contrast the performance of set associative caches, direct.
Lecture 10 Memory Hierarchy and Cache Design Computer Architecture COE 501.
The Memory Hierarchy 21/05/2009Lecture 32_CA&O_Engr Umbreen Sabir.
CS 3410, Spring 2014 Computer Science Cornell University See P&H Chapter: , 5.8, 5.15.
10/18: Lecture topics Memory Hierarchy –Why it works: Locality –Levels in the hierarchy Cache access –Mapping strategies Cache performance Replacement.
CSE 378 Cache Performance1 Performance metrics for caches Basic performance metric: hit ratio h h = Number of memory references that hit in the cache /
Garo Bournoutian and Alex Orailoglu Proceedings of the 45th ACM/IEEE Design Automation Conference (DAC’08) June /10/28.
CSIT 301 (Blum)1 Cache Based in part on Chapter 9 in Computer Architecture (Nicholas Carter)
CPE432 Chapter 5A.1Dr. W. Abu-Sufah, UJ Chapter 5A: Exploiting the Memory Hierarchy, Part 2 Adapted from Slides by Prof. Mary Jane Irwin, Penn State University.
Abdullah Aldahami ( ) March 23, Introduction 2. Background 3. Simulation Techniques a.Experimental Settings b.Model Description c.Methodology.
1 Virtual Memory Main memory can act as a cache for the secondary storage (disk) Advantages: –illusion of having more physical memory –program relocation.
Multiprocessor cache coherence. Caching: terms and definitions cache line, line size, cache size degree of associativity –direct-mapped, set and fully.
1 CMPE 421 Parallel Computer Architecture PART4 Caching with Associativity.
Lecture 5 Cache Operation ECE 463/521 Fall 2002 Edward F. Gehringer Based on notes by Drs. Eric Rotenberg & Tom Conte of NCSU.
Cache Memory By Tom Austin. What is cache memory? A cache is a collection of duplicate data, where the original data is expensive to fetch or compute.
Analyzing Performance Vulnerability due to Resource Denial-Of-Service Attack on Chip Multiprocessors Dong Hyuk WooGeorgia Tech Hsien-Hsin “Sean” LeeGeorgia.
Caches Where is a block placed in a cache? –Three possible answers  three different types AnywhereFully associativeOnly into one block Direct mappedInto.
The Memory Hierarchy Lecture # 30 15/05/2009Lecture 30_CA&O_Engr Umbreen Sabir.
CS2100 Computer Organisation Cache II (AY2015/6) Semester 1.
Influence Of The Cache Size On The Bus Traffic Mohd Azlan bin Hj. Abd Rahman M
M E M O R Y. Computer Performance It depends in large measure on the interface between processor and memory. CPI (or IPC) is affected CPI = Cycles per.
Nov. 15, 2000Systems Architecture II1 Machine Organization (CS 570) Lecture 8: Memory Hierarchy Design * Jeremy R. Johnson Wed. Nov. 15, 2000 *This lecture.
Chapter 9 Memory Organization By Nguyen Chau Topics Hierarchical memory systems Cache memory Associative memory Cache memory with associative mapping.
Cache Memory Chapter 17 S. Dandamudi To be used with S. Dandamudi, “Fundamentals of Computer Organization and Design,” Springer,  S. Dandamudi.
Improving Direct-Mapped Cache Performance by the Addition of a Small Fully-Associative Cache And Pefetch Buffers Norman P. Jouppi Presenter:Shrinivas Narayani.
Review °We would like to have the capacity of disk at the speed of the processor: unfortunately this is not feasible. °So we create a memory hierarchy:
1  2004 Morgan Kaufmann Publishers Chapter Seven Memory Hierarchy-3 by Patterson.
Project 11: Influence of the Number of Processors on the Miss Rate Prepared By: Suhaimi bin Mohd Sukor M
Topics covered: Memory subsystem CSE243: Introduction to Computer Architecture and Hardware/Software Interface.
Princess Sumaya Univ. Computer Engineering Dept. Chapter 5:
Project Summary Fair and High Throughput Cache Partitioning Scheme for CMPs Shibdas Bandyopadhyay Dept of CISE University of Florida.
Lecture 20 Last lecture: Today’s lecture: Types of memory
Cache Memory By Ed Martinez.  The fastest and most expensive memory on a computer system that is used to store collections of data.  Uses very short.
1 Appendix C. Review of Memory Hierarchy Introduction Cache ABCs Cache Performance Write policy Virtual Memory and TLB.
LECTURE 12 Virtual Memory. VIRTUAL MEMORY Just as a cache can provide fast, easy access to recently-used code and data, main memory acts as a “cache”
Constructive Computer Architecture Realistic Memories and Caches Arvind Computer Science & Artificial Intelligence Lab. Massachusetts Institute of Technology.
Recitation 6 – 3/11/01 Outline Cache Organization Replacement Policies MESI Protocol –Cache coherency for multiprocessor systems Anusha
FAMU-FSU College of Engineering 1 Computer Architecture EEL 4713/5764, Fall 2006 Dr. Linda DeBrunner Module #18—Cache Memory Organization.
COMP 3221: Microprocessors and Embedded Systems Lectures 27: Cache Memory - III Lecturer: Hui Wu Session 2, 2005 Modified.
Chapter 9 Memory Organization. 9.1 Hierarchical Memory Systems Figure 9.1.
COSC2410: LAB 19 INTRODUCTION TO MEMORY/CACHE DIRECT MAPPING 1.
1 Memory Hierarchy Design Chapter 5. 2 Cache Systems CPUCache Main Memory Data object transfer Block transfer CPU 400MHz Main Memory 10MHz Bus 66MHz CPU.
Cache Memory Yi-Ning Huang. Principle of Locality Principle of Locality A phenomenon that the recent used memory location is more likely to be used again.
Lecture 12 Virtual Memory.
Replacement Policy Replacement policy:
Set-Associative Cache
Module IV Memory Organization.
Cache - Optimization.
Presentation transcript:

By: Aidahani Binti Ahmad SMP CACHE 2.0 SIMULATOR Project 5 : Influence of the Mapping for Different Cache Sizes. By: Aidahani Binti Ahmad [ M031010031 ]

Design parameters of caches Capacity – total number of blocks the cache can hold. Block size – the number of bytes that are loaded from and written to memory at a time. Associativity – in an N-way set associative cache, a particular block can be loaded in N different cache locations. Replacement policy – which block do we remove from the cache as a new block is loaded. In modern machines, more than one cache is placed between the main memory and the processor.

Project configuration: Configure a system with the following architectural characteristics: • Processors in SMP = 1. • Cache coherence protocol = MESI. • Scheme for bus arbitration = Random. • Word wide (bits) = 32. • Words by block = 64 (block size = 256 bytes). • Blocks in main memory = 4096 (main memory size = 1024 KB). • Replacement policy = LRU. Configure the mapping using : Direct Two-way set associative Four-way set associative Eight-way set associative Fully-associative (remeber: Number_of_ways = Number_of_blocks_in_cache / Number_of_cache_sets).

Project configuration: For each of the configurations of mapping, configure the number of blocks in cache in order to get the following cache sizes: 4 KB (16 blocks in cache) 8 KB (32 blocks in cache) 16 KB (64 blocks in cache) 32 K (128 blocks in cache). For each configuration obtain the miss rate using the memory trace: Ear.

Purpose: Analyze the influence f the mapping on the miss rate for several cache sizes. Simulator Finding: Mapping Cache Size (kb) 4KB 8KB 16KB 32KB 16 block 32 block 64 block 128 block Direct 12.34 7.0271 5.9344 4.0317 2-Way Associative 8.3082 6.7822 3.8809 2.8636 4-Way Associative 6.6315 4.8229 2.939 2.468 8-way Associative 6.5938 5.2562 2.694 2.3361 Full Associative 6.5373 5.1243 2.3962

Does the miss rate increase or decrease as the associativity increases Does the miss rate increase or decrease as the associativity increases? Why? Figure 1: The Miss Rate against mapping type and cache size

Base on the analysis of the finding in Figure 1 it shown that the degree of misses rate decreases when increasing the associative. This because its will be more flexible placement and the conflict misses are reduced. However, 8-way set associative cache has almost the same miss rate as fully associative cache. The most flexible cache system is the fully associative cache where in a fully associative cache subsystem, the caching controller can place a block of bytes in any one of the cache lines present in the cache memory. Greater associativity can come at the great cost.

What does it happen with the conflict misses when you enlarge the associativity grade? Conflict misses means if we use direct-mapped or set-associative strategy two blocks may map to the same record in cache. Figure 2: Miss Rate of associative compare with cache size.

Conflict misses decrease significantly with increased associativity, especially for smaller caches.   Finally, note in Figure 2 that between 16KB there is a large difference of miss rate between direct-mapped and fully associative caches. This difference is the conflict miss rate. The insight from looking at conflict miss rates is that secondary caches benefit a great deal from high associativity.

Does the influence of the associativity grade increase or decrease as the cache size increases? Why?

The associative ways influence decreases with larger caches because the benefits are every time less significant. This conclusion coincides with the theory that conflict misses are reduced by increasing the associativity, however, in large caches this kind of misses are less frequent this can be view in figure 3 where the incluence of associative is very much similar between 8-way associative and full associative.

In conclusion, does the increase of associativity improve the system performance? If the answer is yes, in general, which is the step with more benefits: from direct to 2-way, from 2- way to 4-way, from 4-way to 8-way, or from 8-way to fully-associative?

The system efficiency does improve with increasing the associative ways but the maximum efficiency improvement is – [i] from direct to set- associative with 2 rows in set (2-way).   Cache Rule: the miss rate of a direct-mapped cache of size N is about the same as a two way set-associative cache of size N/2. An 8-way set-associative cache has about the same miss rate as a fully associative cache. A two-way set associative cache is much better than a direct-mapped cache and considerably less complex than a fully associative cache. However, if you're still getting too many conflicts, you might consider using a four-way set associative cache.

Thank You