Home

cachearchitecturen

Cache architectures, or cachearchitecturen in some languages, describe how cache memory is organized and integrated into computing systems to bridge the speed gap between fast processors and slower main memory. They influence performance through hit rate, latency, capacity, energy use, and coherence in multi‑core environments.

In CPUs, caches are typically organized in hierarchical levels, commonly referred to as L1, L2, and L3.

In multi‑core and multi‑processor systems, cache coherence ensures that multiple copies of data in different caches

Replacement policies govern which cache lines are evicted to make room for new data. Common approaches include

Performance is captured by metrics such as cache hit rate, miss penalty, and average memory access time

L1
caches
are
small
and
fast,
often
split
into
separate
instruction
and
data
caches
per
core.
L2
is
larger
and
may
be
private
or
shared,
while
L3
is
usually
larger
still
and
commonly
shared
among
cores.
Caches
can
be
direct-mapped,
fully
associative,
or
set-associative,
with
set
associativity
offering
a
balance
between
speed
and
flexibility.
The
block
size,
or
cache
line,
determines
how
much
data
is
transferred
on
a
miss.
remain
consistent.
Protocols
such
as
MESI
(Modified,
Exclusive,
Shared,
Invalid)
manage
state
transitions.
Inclusion
properties
describe
whether
upper-level
caches
retain
data
that
also
resides
in
lower
levels.
Write
policies
include
write-through
and
write-back,
often
paired
with
write-allocate
or
no-write-allocate
strategies.
least
recently
used
(LRU),
pseudo-LRU,
FIFO,
and
random.
Prefetching
predicts
future
accesses
to
hide
latency
and
improve
hit
rates,
using
patterns,
strides,
or
more
sophisticated
predictors.
(AMAT).
Design
involves
trade-offs
among
latency,
bandwidth,
energy,
area,
and
complexity.
Cache
architectures
extend
beyond
CPUs
to
GPUs,
web
caches,
databases,
and
other
systems,
each
with
domain-specific
requirements
for
coherence,
consistency,
and
eviction.