Skip to main content

Exploratory data analysis tools

Project description

Edamame


Documentation Status PyPI version PyPI - Downloads Maintenance

Edamame is inspired by packages such as pandas-profiling, pycaret, and yellowbrick. The goal of Edamame is to provide user-friendly functions for conducting exploratory data analysis (EDA) on datasets, as well as for training and analyzing batteries of models for regression or classification problems.

To install the package,

pip install edamame

the edamame package works correctly inside a jupyter-notebook. You can find the documentation of the package on the edamame-documentation page.


Functionalities

The package consists of three modules: eda, which performs exploratory data analysis; and regressor and classifier, which handle the training of machine learning models for regression and classification, respectively. To see examples of the uses of the edamame package, you can check out the examples folder in the repository.


Eda module

import edamame.eda as eda

The eda module provides a wide range of functions for performing exploratory data analysis (EDA) on datasets. With this module you can easily explore and manipulate your data, conduct descriptive statistics, correlation analysis, and prepare your data for machine learning. The "eda" module offers the following functionalities:

  • Data Exploration and Manipulation functions:

    • dimensions: The function displays the number of rows and columns of a pandas dataframe passed.
    • identify_types: Identify the data types of each column.
    • view_cardinality: View the number of unique values in each categorical column.
    • modify_cardinality: Modify the number of unique values in a column.
    • missing: Check if any missing data is present in the dataset.
    • handling_missing: Replace or remove missing values in the dataset.
    • drop_columns: Remove specific columns from the dataset.
    • num_to_categorical: The function returns a dataframe with the columns transformed into an "object".
    • interaction: The function display an interactive plot for analysing relationships between numerical columns with a scatterplot.
    • inspection: The function displays an interactive plot for analysing the distribution of a variable based on the distinct cardinalities of the target variable.
    • split_and_scaling: The function returns two pandas dataframes: the regressor matrix X contains all the predictors for the model, the series y contains the values of the response variable.
  • Descriptive Statistics functions:

    • describe_distribution: The function display the result of the describe() method applied to a pandas dataframe, divided by numerical and object columns.
    • plot_categorical: The function returns a sequence of tables and plots for the categorical variables.
    • plot_numerical: The function returns a sequence of tables and plots for the numerical variables.
    • num_variable_study: he function displays the following transformations of the variable col passed: log(x), sqrt(x), x^2, Box-cox, 1/x.
  • Correlation Analysis functions:

    • correlation_pearson: The function performs the Pearson's correlation between the columns pairs.
    • correlation_categorical: The function performs the Chi-Square Test of Independence between categorical variables of the dataset.
    • correlation_phik: Calculate the Phik correlation coefficient between all pairs of columns (Paper link).
  • Useful functions:

    • load_model: The function load the model saved in the pickle format.
    • setup: The function returns the following elements: X_train, y_train, X_test, y_test.
    • scaling: The function returns the normalised/standardized matrix.
    • ohe: The function returns the numpy array passed in as input, converted using one-hot encoding.

Regressor module

from edamame.regressor import TrainRegressor, regression_metrics

The TrainRegressor class is designed to be used as a pipeline for training and handling regression models.

The class provides several methods for fitting different regression models, computing model metrics, saving and loading models, and using AutoML to select the best model based on performance metrics. These methods include:

  • linear: Fits a linear regression model to the training data.
  • lasso: Fits a Lasso regression model to the training data.
  • ridge: Fits a Ridge regression model to the training data.
  • tree: Fits a decision tree regression model to the training data.
  • random_forest: Fits a random forest regression model to the training data.
  • xgboost: Fits an XGBoost regression model to the training data.
  • auto_ml: Uses AutoML to select the best model based on performance metrics.
  • model_metrics: Computes and prints the performance metrics for each trained model.
  • save_model: Saves the trained model to a file.

After saving a model with the save_model method, we can upload the model using the load_model function of the eda module and evaluate its performance on new data using the regression_metrics function.

from edamame.regressor import RegressorDiagnose

The RegressorDiagnose class is designed to diagnose regression models and analyze their performance. The class provides several methods for diagnosing and analyzing the performance of regression models. These methods include:

  • coefficients: Computes and prints the coefficients of the regression model.
  • random_forest_fi: Displays the feature importance plot for the random forest regression model.
  • random_forest_fi: Displays the feature importance plot for the xgboost regression model.
  • prediction_error: Computes and prints the prediction error of the regression model on the test data.
  • residual_plot: creates and displays a residual plot for the regression model.
  • qqplot: creates and displays a QQ plot for the regression model.

Example:

from sklearn.datasets import make_regression
from edamame.regressor import TrainRegressor
import pandas as pd
import edamame.eda as eda
from edamame.regressor import RegressorDiagnose
X, y = make_regression(n_samples=1000, n_features=5, n_targets=1, random_state=42)
X = pd.DataFrame(X, columns=["f1", "f2", "f3", "f4", "f5"])
y = pd.DataFrame(y, columns=["y"])
X_train, y_train, X_test, y_test = eda.setup(X, y)
X_train_s = eda.scaling(X_train)
X_test_s = eda.scaling(X_test)
regressor = TrainRegressor(X_train_s, y_train, X_test_s, y_test)
rf = regressor.random_forest()
regressor.model_metrics()
diagnose = RegressorDiagnose(X_train_s, y_train, X_test_s, y_test)
diagnose.random_forest_fi(model=rf)
diagnose.prediction_error(model=rf)

Classifier module

from edamame.classifier import TrainClassifier

The TrainClassifier class is designed to be used as a pipeline for training and handling clasification models.

The class provides several methods for fitting different regression models, computing model metrics, saving and loading models, and using AutoML to select the best model based on performance metrics. These methods include:

  • logistic: Fits a logistic model to the training data.
  • gaussian_nb: Fits a Gaussina Naive Bayes model to the training data.
  • knn: Fits a k-Nearest Neighbors classification model to the training data.
  • tree: Fits a decision tree classification model to the training data.
  • random_forest: Fits a random forest classification model to the training data.
  • xgboost: Fits an XGBoost classification model to the training data.
    • svm: Fits an Support Vector classification model to the training data.
  • auto_ml: Uses AutoML to select the best model based on performance metrics.
  • model_metrics: Computes and prints the performance metrics for each trained model.
  • save_model: Saves the trained model to a file.

After saving a model with the save_model method, we can upload the model using the load_model function of the eda module and evaluate its performance on new data using the classifier_metrics function.

from edamame.classifier import classifier_metrics

Example:

from edamame.classifier import TrainClassifier
from sklearn import datasets
import edamame.eda as eda
iris = datasets.load_iris()
X = iris.data
X = pd.DataFrame(X, columns=iris.feature_names)
y = iris.target
y = pd.DataFrame(y, columns=['y'])
X_train, y_train, X_test, y_test = eda.setup(X,y)
X_train_s = eda.scaling(X_train)
X_test_s = eda.scaling(X_test)
classifier = TrainClassifier(X_train_s, y_train, X_test_s, y_test)
models = classifier.auto_ml()
svm = classifier.svm()
classifier.model_metrics(model_name="svm")
classifier.save_model(model_name="svm")
svm_upload = eda.load_model(path="svm.pkl")
classifier_metrics(svm_upload, X_train_s, y_train)

Todos

  • Add the notebook for EDA in a classification problem to the edamame-notebook repository.
  • Add the notebook for training/diagnosing classification models to the edamame-notebook repository.
  • Add the rocauc method to the ClassifierDiagnose class.
  • Update example notebooks.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

edamame-0.59.tar.gz (29.0 kB view details)

Uploaded Source

Built Distribution

edamame-0.59-py3-none-any.whl (29.4 kB view details)

Uploaded Python 3

File details

Details for the file edamame-0.59.tar.gz.

File metadata

  • Download URL: edamame-0.59.tar.gz
  • Upload date:
  • Size: 29.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.12

File hashes

Hashes for edamame-0.59.tar.gz
Algorithm Hash digest
SHA256 cf41505ae3936bb18ec671974f7e9453274d0588ca7a794b9f4d2bc6d43177e1
MD5 ddbfbea3f23c62a863f35c428db5bb8a
BLAKE2b-256 fa6e0a28342aa3952bb6dd6229016cf9b2fb3473152ea71652144ac89858a10e

See more details on using hashes here.

File details

Details for the file edamame-0.59-py3-none-any.whl.

File metadata

  • Download URL: edamame-0.59-py3-none-any.whl
  • Upload date:
  • Size: 29.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.12

File hashes

Hashes for edamame-0.59-py3-none-any.whl
Algorithm Hash digest
SHA256 48c2e202535da04ce87f403ed0e0db44e691fcfeca309c7589f2f27f11fc604f
MD5 bd5300db5dba87455fbe016b6d7e0a9b
BLAKE2b-256 8bc47f41a8e583e9733e4904999971c047a4357d9c1251a433e281dd11479f4d

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page