Advanced regression methods with sklearn-like interface
Project description
Better Regressions
Advanced regression methods with an sklearn-like interface.
Current Features
Linear:- Configurable regularization: Ridge with given
alpha/ BayesianRidge / ARD - "Better bias" option to properly regularize the intercept term
- Configurable regularization: Ridge with given
Scaler:- Configurable preprocessing: Standard scaling (by second moment) / Quantile transformation with uniform/normal output / Power transformation
AutoScalerto automatically select the best scaling method based on validation split
Smooth: Boosting-based regression using smooth functions for featuresSuperSmoother: Adaptive-span smoother for arbitrary complex functions.Angle: Bagging of piecewise-linear functions, it's less flexible but because of that it's more robust to overfitting.
Soft: Mixture of regressors based on quantile classificationStabilize: Robust scaling & clipping transformation for features/targetsAutoClassifier: Classification with automatic model selection (LogisticRegression or XGBoost, with auto depth selection)Adapter: Bins features and target, then trains a classifier. This way it can learn non-linear relationships and it also models the target distribution (not only its mean).EDA: Exploratory Data Analysis utilitiesplot_distribution: Visualize sample distributions with fitted t-distribution parametersplot_trend: Automatically detect and visualize relationships between variables + Pearson/Spearman correlation- For discrete features: Shows violin plots with distribution at each value
- For continuous features: Fits trend lines with variance estimation and confidence intervals
Installation
pip install better-regressions
Basic Usage
from better_regressions import auto_angle, auto_linear, Linear, Scaler, AutoClassifier
from better_regressions.eda import plot_distribution, plot_trend
from sklearn.datasets import make_regression, make_moons
import numpy as np
X, y = make_regression(n_samples=100, n_features=5, noise=0.1)
model = auto_angle(n_breakpoints=2)
model.fit(X, y)
y_pred = model.predict(X)
print(repr(model))
# Classification example
dataset = make_moons(n_samples=200, noise=0.3)
Xc, yc = dataset
clf = AutoClassifier(depth="auto")
clf.fit(Xc, yc)
yc_pred = clf.predict(Xc)
# EDA example
plot_distribution(y, name="Target Distribution")
plot_trend(X[:, 0], y, name="Feature 0 vs Target")
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
better_regressions-0.5.0.tar.gz
(447.9 kB
view details)
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file better_regressions-0.5.0.tar.gz.
File metadata
- Download URL: better_regressions-0.5.0.tar.gz
- Upload date:
- Size: 447.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.11.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
8ad7313f1cea1da332451b87273377d5b967ffca0144df0bc89d1cd9d3dafb2e
|
|
| MD5 |
f6654b977619a260153ff5187ee7c3b1
|
|
| BLAKE2b-256 |
f046b438c7c62ad772517c0034846745b99f4426f7d0d325774be80fbeb07ca7
|
File details
Details for the file better_regressions-0.5.0-py3-none-any.whl.
File metadata
- Download URL: better_regressions-0.5.0-py3-none-any.whl
- Upload date:
- Size: 19.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.11.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
e52e88c9ea0dddbae35e26681906e72f98831083b6d73b4c1a2566f7ac05c067
|
|
| MD5 |
8c2735e9c6832cef8ac42b581d494412
|
|
| BLAKE2b-256 |
5eee3f4332a06ab2dd1ab8adba73772f084d6f098f5af88236c88e7540c03a54
|