Skip to main content

Python Cache Hierarchy Simulator

Project description

pycachesim

A single-core cache hierarchy simulator written in python.

https://travis-ci.org/RRZE-HPC/pycachesim.svg?branch=master

The goal is to accurately simulate the caching (allocation/hit/miss/replace/evict) behavior of all cache levels found in modern processors. It is developed as a backend to kerncraft, but is also planned to introduce a command line interface to replay LOAD/STORE instructions.

Currently supported features:
  • Inclusive cache hierarchies

  • LRU, MRU, RR and FIFO policies

  • N-way cache associativity

  • Write-allocate with write-back caches

  • Non-write-allocate with write-through caches

  • Write-combining with sub-blocking

  • Tracking of cacheline states (e.g., using dirty bits)

  • Speed (core is implemented in C)

  • Python 2.7+ and 3.4+ support, with no other dependencies

Planned features:
  • Report cachelines on all levels (preliminary support through backend.verbosity > 0)

  • Report timeline of cache events (preliminary support through backend.verbosity > 0)

  • Visualize events (html file?)

  • Interface to Valgrind Infrastructure (see Lackey) for access history replay.

  • (uncertain) instruction cache

  • Optional classification into compulsory/capacity and conflict misses (by simulating other cache configurations in parallel)

  • (uncertain) multi-core support

License

pycachesim is licensed under AGPLv3.

Usage

from cachesim import CacheSimulator, Cache, MainMemory

mem = MainMemory()
l3 = Cache("L3", 20480, 16, 64, "LRU")  # 20MB: 20480 sets, 16-ways with cacheline size of 64 bytes
mem.load_to(l3)
mem.store_from(l3)
l2 = Cache("L2", 512, 8, 64, "LRU", store_to=l3, load_from=l3)  # 256KB
l1 = Cache("L1", 64, 8, 64, "LRU", store_to=l2, load_from=l2)  # 32KB
cs = CacheSimulator(l1, mem)

cs.load(2342)  # Loads one byte from address 2342, should be a miss in all cache-levels
cs.store(512, length=8)  # Stores 8 bytes to addresses 512-519,
                         # will also be a load miss (due to write-allocate)
cs.load(512, length=8)  # Loads from address 512 until (exclusive) 520 (eight bytes)

cs.force_write_back()
cs.print_stats()

This should return:

CACHE *******HIT******** *******MISS******* *******LOAD******* ******STORE*******
   L1      1 (       8B)      2 (      65B)      3 (      73B)      1 (       8B)
   L2      0 (       0B)      2 (     128B)      2 (     128B)      1 (      64B)
   L3      0 (       0B)      2 (     128B)      2 (     128B)      1 (      64B)
  MEM      2 (     128B)      0 (       0B)      2 (     128B)      1 (      64B)

Each row refers to one memory-level, starting with L1 and ending with main memory. The 3 loads in L1 are the sum of all individual accesses to the cache-hierarchy. 1 (from first load) + 1 (from store with write-allocate) + 1 (from second load) = 3.

The 1 hit, is for bytes which were cached already. Internally the pycachesim operates on cache-lines, which all addresses get transformed to. Thus, the two misses throughout all cache-levels are actually two complete cache-lines and after the cache-line had been loaded the consecutive access to the same cache-line are handled as hits. That is also the reason why data sizes increase from L1 to L2. L1 is accessed byte-wise and L2 only with cache-line granularity.

So: hits, misses, stores and loads in L1 are byte-wise. Every other statistical information are based on cache-lines.

When using victim caches, setting victims_to to the victim cache level, will cause pycachesim to forward unmodified cache-lines to this level on replacement. During a miss, victims_to is checked for availability and only hit if it the cache-line is found. This means, that load stats will equal hit stats in victim caches and misses should always be zero.

Comparison to other Cache Simulators

While searching for more versatile cache simulator for kerncraft, I stumbled across the following:

  • gem5: Very fully-featured full system simulator. Complex to extract only the memory subsystem

  • dineroIV: Nice and simple code, but does not support exclusive caches and not available under open source license.

  • cachegrind: Maintained and stable code of a well established open source project, but only supports inclusive first and last level caches.

  • callgrind: see cachegrind

  • SMPcache: Only supports one single cache and runs on Windows with GUI. Also not freely available.

  • CMPsim: Was only academically published and source code never made available.

  • CASPER: Was only academically published and source code never made available.

Package

instructions [0]

blocks [1]

sub-blocks [2]

associtivity [3]

LRU [4]

MRU [4]

FIFO [4]

RR [4]

CCC [5]

3+ levels [6]

exclusive [7]

victim [8]

multi-core [9]

API [10]

open source [11]

gem5

x

x

?

x

x

x

x

?

?

x

?

?

?

python, ruby, c++

yes, BSD-style

dineroIV

x

x

x

x

x

x

x

x

x

c

no, free for non-comercial use

cachegrind

x

x

x

x

cli

yes, GPLv2

callgrind

x

x

x

x

cli

yes, GPLv2

SMPcache

x

x

x

x

x

?

Windows GUI

no, free for education und research

CMPsim

x

x

x

x

x

x

x

?

?

x

?

no, source not public

CASPER

x

x

x

x

x

x

x

x

x

x

x

perl, c

no, source not public

pycachesim

x

x

x

x

x

x

x

x

x

x

python, C backend

yes, AGPLv3

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pycachesim-0.1.8.tar.gz (20.5 kB view details)

Uploaded Source

File details

Details for the file pycachesim-0.1.8.tar.gz.

File metadata

  • Download URL: pycachesim-0.1.8.tar.gz
  • Upload date:
  • Size: 20.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.12.1 pkginfo/1.4.2 requests/2.20.0 setuptools/40.4.3 requests-toolbelt/0.8.0 tqdm/4.28.1 CPython/2.7.14

File hashes

Hashes for pycachesim-0.1.8.tar.gz
Algorithm Hash digest
SHA256 8e24b946d95f9148d4af51e25404a1a53f255a07c91a8a5b8b3273942f0282c6
MD5 147b5250d59d44a790dc512a8e33c44a
BLAKE2b-256 65e354e6eb42a2a4b4fedadbd745432367adb69cc9edcc4775b58bc73afe6f1e

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page