体系结构量化研究方法(第二章-2)

Chapter 2 Memory Hierarchy Design

Memory Technology and Optimizations

Performance metrics

  • Latency
    • Access time
      • Time between read request and when desired word arrives
    • Cycle time
      • Minimum time between unrelated requests to memory
    • Bandwidth

![[Pasted image 20241220174926.png|500]]

High Bandwidth Memory (HBM)

  • A packaging innovation rather than a circuit innovation
  • Reduce access latency by shortening the delay between the DRAM and the processor
  • Interposer stacking (2.5D) available, vertical stacking (3D) still under development due to heat constraint

![[Pasted image 20241220175322.png|650]]

Six Basic Cache Optimizations

Average memory access time:

  • Average memory access time = Miss rate × Miss penalty + Hit Time

Six basic cache optimizations:

  • Reduce the miss rate
    • Larger block size (reduce compulsory misses)
      • Increase conflict misses, Increases miss penalty
    • Larger total cache capacity (reduce capacity misses)
      • Increase hit time, Increase cost, Increase power
    • Higher associativity (reduce conflict misses)
      • Increase hit time
        • More tag comparisons
      • Increase power
  • Reduce the miss penalty
    • Higher number of cache levels
      • L1 miss penalty can be reduced by introducing L2 cache
      • balance fast-hits and few-misses
    • Giving priority to read misses over writes
      • “read-after-write” data hazard
        • Let read wait until write buffer is empty
        • Or check the content of the write buffer on a read miss
          • conflict -> wait
          • non-conflict -> handle read miss first
        • Prioritizing read can reduce unnecessary stalling of read miss
  • Reduce the time to hit in cache
    • Avoiding address translation in cache indexing
      • Virtual Cache avoids address translation
        • virtual address -> cached block
          • use virtual address in both set indexing & tag comparison
      • Obstacles to realize virtual cache
        • To ensure protection
          • copy protection bit of TLB into an added field on cache miss
        • To support process switching
          • adopt PID marker in cache
        • To ensure compatibility with multiple alias logic addresses (cache consistency)
          • antialiasing: compare physical addresses of existing blocks with the fetched one

![[Pasted image 20241220180510.png|650]]

![[Pasted image 20241220175802.png]]

Ten Advanced Cache Optimizations

Advanced Cache Optimizations

  • Reducing the hit time
  • Reducing the miss penalty
  • Reducing the miss rate
  • Increasing cache bandwidth
  • Reducing th
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值