Skip to main content

Use MCMC to do benchmark analysis

Project description

benchmcmc — Benchmark analysis with MCMC

Install

Installing is as always

pip install benchmcmc

It depends on pymc3 and matplotlib, so these are also installed.

Quickstart

Getting started real quick:

benchmcmc --generate 69 11 10 131 10 9 --beta > bench.txt
benchmcmc bench.txt

The first command generates 200 synthetic data points and put them into bench.txt, and the second command runs benchmcmc on the data, displaying a trace plot of the MCMC.

Introduction

This package lets you take a series of benchmark data analyses whether there at one point was a change in performance.

Suppose that you run your benchmark tests on every commit that you have (e.g. looping over git-rev-list), and you see that your performance data is (e.g. in requests per second or in seconds, or other measures) as follows:

13.64
12.82
11.69
15.12
12.30
18.46
13.51
14.33
13.84
12.77
... (180 rows omitted)
10.93
11.02
12.45
11.78
12.12
13.51
10.66
10.18
10.81
12.19

In this data, it seems to be centered around 13+ε in the beginning, and it ends being centered around 12+ε, or visualized:

scatterplot of performance over time

There seems to be a slight drop in values before the 100th point, but it's not easy to determine exactly where the change occurred.

Suppose that you wonder whether or not the performance at the start and at the end are likely to be from two different distributions, and if so, where the switchpoint was.

traceplot

Analysis

Running benchmcmc on the data gives the above plot which shows that it is likely that the performance went from ~13.5 to ~12.25 at or around the 69th or 75th datapoints.

This helps you pin down when a performance change might have occurred.

Usage

$ benchmcmc bench.txt --draws 1000 --tune 1000 --cores 4 --target-accept 0.8

The options available are those illustrated in the above examples and all taken from and fed into pymc3.sampling.sample.

  • draws — The number of samples to draw. Defaults to 1000. The number of tuned samples are discarded by default.
  • tune — Number of iterations to tune, defaults to 1000. Samplers adjust the step sizes, scalings or similar during tuning. Tuning samples will be drawn in addition to the number specified in the draws argument, and will be discarded.
  • cores — The number of chains to run in parallel. If None, set to the number of CPUs in the system, but at most 4.
  • target_accept : float in [0, 1] — The step size is tuned such that we approximate this acceptance rate. Higher values like 0.9 or 0.95 often work better for problematic posteriors.

Generating synthetic data

You can run benchmcmc --generate for generating synthetic benchmark data.

$ benchmcmc --generate 100 15 3 100 14 3 [--beta] > benchmarkfile.txt

This generates 200 samples, 100 from N(mu=15, sigma=3) followed by 100 from N(mu=14, sigma=3).

If you use --beta, you get a bit more realistic performance with a lower bound of mu, especially for lower values of mu.

Running a script on a history

Suppose that you want to run python script.py on a script that is in your Git tree.

LOGFILE=/tmp/timescript
echo "" > $LOGFILE
for commit in $(git rev-list master)
do
    git checkout $commit
    printf "%s," "`(git rev-parse --short HEAD)`" >> $LOGFILE
    /usr/bin/time -a -o $LOGFILE --format=%e python script.py
done
tac $LOGFILE

When run in a repository, it will output time data in the format

commit,time

Here is an example of the output:

484fde8,0.04
58a1cdb,0.04
d26b797,0.04
81f4b9a,0.04
3ae1e11,0.04
7689ca2,0.04
8c76b29,0.04
43db50c,0.04
b34b146,0.04
4c56a54,0.04
9c08050,0.07
b22278d,0.07
7a9c111,0.07
065b6a5,0.07
6cc7cdd,0.07
ec7f042,0.07
b3ba887,0.08
a32ce81,0.07
9136914,0.07
b456714,0.07
504cf73,0.07
8002774,0.07
e1f5f9f,0.09

For example, you can create a script gitrun that looks as follows:

#!/usr/bin/bash

GITRUN_EXEC=$1
shift
GITRUN_ARGS=$@

for commit in $(git rev-list master)
do
  git checkout --quiet $commit
  printf "%s," "`(git rev-parse --short HEAD)`"
  sh -c "$GITRUN_EXEC $GITRUN_ARGS"
done

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

benchmcmc-0.0.7.tar.gz (5.9 kB view details)

Uploaded Source

Built Distribution

benchmcmc-0.0.7-py3-none-any.whl (6.8 kB view details)

Uploaded Python 3

File details

Details for the file benchmcmc-0.0.7.tar.gz.

File metadata

  • Download URL: benchmcmc-0.0.7.tar.gz
  • Upload date:
  • Size: 5.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.6.1 requests/2.25.0 setuptools/51.0.0 requests-toolbelt/0.9.1 tqdm/4.54.0 CPython/3.6.9

File hashes

Hashes for benchmcmc-0.0.7.tar.gz
Algorithm Hash digest
SHA256 ee8c8b2156fb73afba4fb2e8d071357ab91e079eb932a1a62bfc52d57a047454
MD5 d25cf4e5d1bb25dd7a0b972e3b9def0f
BLAKE2b-256 3374bd4e0ebbdc1957bb9d329296f92e586cb493d1fb40cf4704dae256dc31f9

See more details on using hashes here.

File details

Details for the file benchmcmc-0.0.7-py3-none-any.whl.

File metadata

  • Download URL: benchmcmc-0.0.7-py3-none-any.whl
  • Upload date:
  • Size: 6.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.6.1 requests/2.25.0 setuptools/51.0.0 requests-toolbelt/0.9.1 tqdm/4.54.0 CPython/3.6.9

File hashes

Hashes for benchmcmc-0.0.7-py3-none-any.whl
Algorithm Hash digest
SHA256 0cc1017c5a387a70cf4acc29c963f92f0ac74f67a994eb24a433667128a7fe97
MD5 4e337bf32f5575a3c1d3dcd406ea5fc3
BLAKE2b-256 783369053c2864cf1114d2a440118fc6c824740651d939b37343a5d2bed11460

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page