CAQA_Memory Flashcards

(65 cards)

1
Q

This ensures that nearly all references can be found in smaller memories.

A

Temporal and spatial locality

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

This gives the allusion of a large, fast memory being presented to the processor

A

Temporal and Spatial Locality

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

In a memory hierarchy diagram, the _________ are put at the top while the ______ are put at the bottom

A

faster and smaller, slower and larger

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Why is the memory hierarchy design crucial in recent multi-core processors?

A

CPU speed increases much faster memory speed, leading to the memory wall. Caches are needed to close this gap.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does CPUs need to cope with its huge bandwidth demand?

A

multi-port caches
per-core L1/L2 caches
shared L3 cache

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

High-end microprocessors have _____ on chip cache.

A

> 10 MB

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

When a word/data is not found in the cache, a __________ occurs.

A

miss

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does the cache do when a miss occurs?

A

fetch from lower level in hierarchy that may be another cache or memory. Fetching other words within the block taking advantage of spatial locality

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

n-way = __________

A

n-blocks per set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

If one block per set, the cache is

A

direct-mapped (1-way)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

If the cache is 1 way but with all blocks inside it, its associativity is

A

fully associative

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Immediately update lower levels of hierarchy

A

Write-through

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Only update lower levels of hierarchy when an updated block is replaced

A

Write-back

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Writing strategies use ___________ to make writes asynchronous.

A

write buffer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Fraction of cache access that result in a miss

A

Miss rate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

cause of a miss: first reference to a block

A

Compulsory

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

cause of miss that blocks discarded and later retrieved

A

capacity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

cause of miss: program makes repeated references to multiple addresses from different blocks that map to the same location in the cache

A

conflict

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Average memory access time =

A

Hit time + Miss rate X Miss Penalty

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

misses per instruction =

A

miss rate X memory accesses/instruction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

To reduce performance impact of misses

A

speculative and multithreaded processors may execute other instructions during miss

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Reduces compulsory misses and Increases capacity and conflict misses, increases miss penalty

A

Larger block size

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Increases hit time, increases power consumption

A

Larger total cache capacity to reduce miss rate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Reduces conflict misses and increases hit time, power consumption

A

Higher associativity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Reduces overall memory access time
Higher number of cache levels
26
Reduces miss penalty
Giving priority to read misses over writes
27
Reduces hit time
Avoiding address translation in cache indexing
28
Time between read request and when desired word arrives
Access time
29
Minimum time between unrelated requests to memory
Cycle time
30
Has low latency, use for cache
SRAM
31
Chips into many banks for high bandwidth, use for main memory
Organize DRAM
32
Requires low power to retain bit and 6 transistors/bit
SRAM
33
Must be re-written after being read and must also be periodically refreshed
DRAM
34
In DRAM, upper half of address is
row access strobe (RAS)
35
In DRAM, lower half of address is
column access strobe (CAS)
36
Some optimizations in memory capacity and speed to keep up with processors.
Multiple accesses to same row Synchronous DRAM Wider interfaces Double data rate (DDR) Multiple banks on each DRAM device
37
DDR with lower power (2.5V -> 1.8 V) and clock rates (266 -> 400 MHz)
DDR2
38
DDR with 1.5 V power and 800 MHz clock rate
DDR3
39
DDR with 1-1.2 V power and 1333 MHz clock rate
DDR4
40
Graphics memory based on DDR3
GDDR5
41
Has lower voltage and lower power mode (ignores clock, continues refresh)
SDRAMs
42
Achieve 2-5 X bandwidth per DRAM vs DDR3 and has wider interfaces and higher clock rate
Graphics Memory
43
DRAM stacked vertically with very high bandwidth and used in GPUs and HPC
High Bandwidth Memory (HBM)
44
Non-volatile that is slower than DRAM but faster than disk. It has two types (NAND and NOR)
Flash Memory
45
Must be erased in blocks before being overwritten and can use as little as zero power.
NAND Flash Memory
46
NAND Flash Memory is limited to _____________- number of cycles
100,00
47
Memory is susceptible to __________.
cosmic rays/ soft errors
48
Soft errors are fixed by _____
ECC
49
Hard Errors are permanent errors that are fixed by
replacing it by spare rows
50
a RAID-like error recovery technique
Chipkill
51
To improve hit time, predict the _____ to pre-set mux
way
52
To improve bandwidth
Pipelined caches
53
To support simultaneous access
Multibanked Caches
54
Allow hits while misses are outstanding
Nonblocking caches
55
Request missed word from memory first and send it to the processor as soon as it arrives
Critical word first
56
Request words in normal order and send missed work to the processor as soon as it arrives
Early restart
57
Swaps nested loops to access memory in sequential order
Loop interchange
58
Improves locality of accesses
Blocking
59
Fetch two blocks on miss (include next sequential block)
Hardware Prefetching
60
Insert prefetch instructions before data is needed
Compiler prefetching
61
HBM as a cache where each SDRAM row is a block index and contains set of tags and 29 segments. Hit requires a CAS.
L-H
62
HBM as cache that molds tag and data together and use direct mapped
Alloy cache
63
Keeps processes in their own memory space
Virtual Memory
64
Provide user mode and supervisor mode Protect certain aspects of CPU state Provide mechanisms for switching between user mode and supervisor mode Provide mechanisms to limit memory accesses Provide TLB to translate addresses
Role of architecture
65
Supports isolation and security. Sharing a computer among many unrelated users Enabled by raw speed of processors, making the overhead more acceptable
Virtual Machine