Very easy Bayesian regression using numpyro.
Project description
Shabadoo: very easy Bayesian regression.
"That's the worst name I ever heard."
Shabadoo is the worst kind of machine learning. It automates nothing; your models will not perform well and it will be your own fault.
Shabadoo is for people who want to do Bayesian regression but who do not want to write probabilistic programming code. You only need to assign priors to features and pass your pandas dataframe to a .fit()
/ .predict()
API.
Shabadoo runs on numpyro and is basically a wrapper around the numpyro Bayesian regression tutorial.
Quickstart
Install
pip install shabadoo
or
pip install git+https://github.com/nolanbconaway/shabadoo
Specifying a Shabadoo Bayesian model
Shabadoo was designed to make it as easy as possible to test ideas about features and their priors. Models are defined using a class which contains configuration specifying how the model should behave.
You need to define a new class which inherits from one of the Shabadoo models. Currently, Normal, Poisson, and Bernoulli are implemented.
import pandas as pd
from numpyro import distributions as dist
from shabadoo import Normal
# fake data
df = pd.DataFrame(dict(x=[1, 2, 2, 3, 4, 5], y=[1, 2, 3, 4, 3, 5]))
class Model(Normal):
dv = "y"
features = dict(
const=dict(transformer=1, prior=dist.Normal(0, 1)),
x=dict(transformer=lambda df: df.x, prior=dist.Normal(0, 1)),
)
The dv
attribute specifies the variable you are predicting. features
is a dictionary of dictionaries, with one item per feature. Above, two features are defined (const
and x
). Each feature needs a transformer
and a prior
.
The transformer specifies how to obtain the feature given a source dataframe. The prior specifies your beliefs about the model's coefficient for that feature.
Fitting & predicting the model
Shabadoo models implement the well-known .fit
/ .predict
api pattern.
model = Model().fit(df)
# sample: 100%|██████████| 1500/1500 [00:05<00:00, 282.76it/s, 7 steps of size 4.17e-01. acc. prob=0.88]
model.predict(df)
"""
0 1.309280
1 2.176555
2 2.176555
3 3.043831
4 3.911106
5 4.778381
"""
Use model.predict(df, ci=True)
to obtain a confidence interval around the model's prediction.
Inspecting the model
Shabadoo's model classes come with a number of model inspection methods. It should be easy to understand your model's composition and with Shabadoo it is!
Print the model formula
The average and standard deviation of the MCMC samples are used to provide a rough sense of the coefficient in general.
print(model.formula)
"""
y = (
const * 0.44200(+-0.63186)
+ x * 0.86728(+-0.22604)
)
"""
Measure prediction accuracy.
The Model.metrics()
method is packed with functionality. You should not have to write a lot of code to evaluate your model's prediction accuracy!
Obtaining aggregate statistics is as easy as:
model.metrics(df)
{'r': 0.8646920305474705,
'rsq': 0.7476923076923075,
'mae': 0.5663623639121652,
'mape': 0.20985123644135573}
For per-point errors, use aggerrs=False
. A pandas dataframe will be returned that you can join on your source data using its index.
model.metrics(df, aggerrs=False)
"""
residual pe ape
0 -0.309280 -30.928012 30.928012
1 -0.176555 -8.827769 8.827769
2 0.823445 27.448154 27.448154
3 0.956169 23.904233 23.904233
4 -0.911106 -30.370198 30.370198
5 0.221619 4.432376 4.432376
"""
You can use grouped_metrics
to understand within-group errors. Under the hood, the predicted and actual dv
are groupby-aggregated (default sum) and metrics are computed within each group.
df["group"] = [1, 1, 1, 2, 2, 2]
model.grouped_metrics(df, 'group')
{'r': 1.0, 'rsq': 1.0, 'mae': 0.30214565559127315, 'mape': 0.03924585080786096}
model.grouped_metrics(df, "group", aggerrs=False)
"""
residual pe ape
group
1 -0.337609 -5.626818 5.626818
2 -0.266682 -2.222352 2.222352
"""
Saving and recovering a saved model
Shabadoo models have a from_samples
method which allows a model to be save and recovered exactly.
Samples from fitted models can be accessed using model.samples
and model.samples_df
.
model.samples['x']
"""
DeviceArray([0.65721655, 0.7644873 , 0.8724553 , 0.6285299 , 0.681262 ,
...
"""
model.samples_df.head()
"""
const x
0 0.689248 0.657217
1 0.524834 0.764487
2 1.093962 0.872455
3 1.253354 0.628530
4 1.021025 0.681262
"""
Use the samples to recover your model:
model_recovered = Model.from_samples(model.samples)
model_recovered.predict(df).equals(model.predict(df))
True
Model samples can be saved as JSON using model.samples_json
:
import json
with open('model.json', 'w') as f:
f.write(model.samples_json)
with open('model.json', 'r') as f:
model_recovered = Model.from_samples(json.load(f))
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file shabadoo-0.0.3.tar.gz
.
File metadata
- Download URL: shabadoo-0.0.3.tar.gz
- Upload date:
- Size: 13.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/45.1.0 requests-toolbelt/0.9.1 tqdm/4.43.0 CPython/3.7.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 08608bacd78b555c1473c6279033e783bc22522de9e800ae81b17369f3167528 |
|
MD5 | 3f1e6cc4e65f1e56823a8660aba497e0 |
|
BLAKE2b-256 | 3293f4c8aea08e097bdc12be3942ee15c1c777254c9c934c56de78bb1c1ee2ee |
File details
Details for the file shabadoo-0.0.3-py3-none-any.whl
.
File metadata
- Download URL: shabadoo-0.0.3-py3-none-any.whl
- Upload date:
- Size: 9.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/45.1.0 requests-toolbelt/0.9.1 tqdm/4.43.0 CPython/3.7.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 351f02341c9dfae63de26a6c9e3cc9c77473da5c3968d63334eb0a9c86ea2e41 |
|
MD5 | d4f1a0006ae56b31649aa4e2b2d58a6f |
|
BLAKE2b-256 | 923eaa59981593df1ecfe0cfd9c2e1b7d9718b802967c3fa77cdfd35732cd155 |