Skip to main content

This project is an ensemble of methods which are frequently used in python Data Science projects.

Project description

Data Science Utils: Frequently Used Methods for Data Science

License: MIT GitHub release (latest SemVer) GitHub issues Documentation Status PyPI - Python Version PyPI - Wheel PyPI version Anaconda-Server Badge Build Status Coverage Status

Data Science Utils extends the Scikit-Learn API and Matplotlib API to provide simple methods that simplify tasks and visualizations for data science projects.

Code Examples and Documentation

Let's explore some code examples and outputs.

You can find the full documentation with all the code examples at: https://datascienceutils.readthedocs.io/en/latest/

In the documentation, you can find more methods and additional examples.

The API of the package is built to work with the Scikit-Learn API and Matplotlib API. Here are some of the capabilities of this package:

Metrics

Plot Confusion Matrix

Computes and plots a confusion matrix, False Positive Rate, False Negative Rate, Accuracy, and F1 score of a classification.

from ds_utils.metrics import plot_confusion_matrix

plot_confusion_matrix(y_test, y_pred, [0, 1, 2])

multi label classification confusion matrix

Plot Metric Growth per Labeled Instances

Receives train and test sets, and plots the given metric change with an increasing number of trained instances.

from ds_utils.metrics import plot_metric_growth_per_labeled_instances
from sklearn.tree import DecisionTreeClassifier
from sklearn.ensemble import RandomForestClassifier

plot_metric_growth_per_labeled_instances(
    x_train, y_train, x_test, y_test,
    {
        "DecisionTreeClassifier": DecisionTreeClassifier(random_state=0),
        "RandomForestClassifier": RandomForestClassifier(random_state=0, n_estimators=5)
    }
)

metric growth per labeled instances with n samples

Visualize Accuracy Grouped by Probability

Receives test true labels and classifier probability predictions, divides and classifies the results, and finally plots a stacked bar chart with the results. Original code

from ds_utils.metrics import visualize_accuracy_grouped_by_probability

visualize_accuracy_grouped_by_probability(
    test["target"],
    1,
    classifier.predict_proba(test[selected_features]),
    display_breakdown=False
)

Without breakdown:

visualize_accuracy_grouped_by_probability

With breakdown:

visualize_accuracy_grouped_by_probability_with_breakdown

Receiver Operating Characteristic (ROC) Curve with Probabilities (Thresholds) Annotations

Plot ROC curves with threshold annotations for multiple classifiers, using plotly as a backend.

from ds_utils.metrics import plot_roc_curve_with_thresholds_annotations

classifiers_names_and_scores_dict = {
    "Decision Tree": tree_clf.predict_proba(X_test)[:, 1],
    "Random Forest": rf_clf.predict_proba(X_test)[:, 1],
    "XGBoost": xgb_clf.predict_proba(X_test)[:, 1]
}
fig = plot_roc_curve_with_thresholds_annotations(
    y_true,
    classifiers_names_and_scores_dict,
    positive_label=1
)
fig.show()

plot roc curve with thresholds annotations

Precision-Recall Curve with Probabilities (Thresholds) Annotations

Plot Precision-Recall curves with threshold annotations for multiple classifiers, using plotly as a backend.

from ds_utils.metrics import plot_precision_recall_curve_with_thresholds_annotations

classifiers_names_and_scores_dict = {
    "Decision Tree": tree_clf.predict_proba(X_test)[:, 1],
    "Random Forest": rf_clf.predict_proba(X_test)[:, 1],
    "XGBoost": xgb_clf.predict_proba(X_test)[:, 1]
}
fig = plot_precision_recall_curve_with_thresholds_annotations(
    y_true,
    classifiers_names_and_scores_dict,
    positive_label=1
)
fig.show()

plot precision recall curve with thresholds annotations

Preprocess

Visualize Feature

Receives a feature and visualizes its values on a graph:

  • If the feature is float, the method plots the distribution plot.
  • If the feature is datetime, the method plots a line plot of progression over time.
  • If the feature is object, categorical, boolean, or integer, the method plots a count plot (histogram).
from ds_utils.preprocess import visualize_feature

visualize_feature(X_train["feature"])
Feature Type Plot
Float
Integer
Datetime
Category / Object
Boolean

Get Correlated Features

Calculate which features are correlated above a threshold and extract a data frame with the correlations and correlation to the target feature.

from ds_utils.preprocess import get_correlated_features

correlations = get_correlated_features(train, features, target)
level_0 level_1 level_0_level_1_corr level_0_target_corr level_1_target_corr
income_category_Low income_category_Medium 1.0 0.1182165609358650 0.11821656093586504
term_ 36 months term_ 60 months 1.0 0.1182165609358650 0.11821656093586504
interest_payments_High interest_payments_Low 1.0 0.1182165609358650 0.11821656093586504

Visualize Correlations

Compute pairwise correlation of columns, excluding NA/null values, and visualize it with a heat map. Original code

from ds_utils.preprocess import visualize_correlations

visualize_correlations(data)

visualize features

Plot Correlation Dendrogram

Plot a dendrogram of a correlation matrix. This consists of a chart that shows hierarchically the variables that are most correlated by connecting trees. The closer to the right that the connection is, the more correlated the features are. Original code

from ds_utils.preprocess import plot_correlation_dendrogram

plot_correlation_dendrogram(data)

plot correlation dendrogram

Plot Features' Interaction

Plots the joint distribution between two features:

  • If both features are either categorical, boolean, or object, the method plots the shared histogram.
  • If one feature is either categorical, boolean, or object and the other is numeric, the method plots a boxplot chart.
  • If one feature is datetime and the other is numeric or datetime, the method plots a line plot graph.
  • If one feature is datetime and the other is either categorical, boolean, or object, the method plots a violin plot ( combination of boxplot and kernel density estimate).
  • If both features are numeric, the method plots a scatter graph.
from ds_utils.preprocess import plot_features_interaction

plot_features_interaction("feature_1", "feature_2", data)
Numeric Categorical Boolean Datetime
Numeric
Categorical
Boolean
Datetime

Strings

Append Tags to Frame

This method extracts tags from a given field and appends them as new columns to the dataframe.

Consider a dataset that looks like this:

x_train:

article_name article_tags
1 ds,ml,dl
2 ds,ml

x_test:

article_name article_tags
3 ds,ml,py

Using this code:

import pandas as pd
from ds_utils.strings import append_tags_to_frame

x_train = pd.DataFrame([{"article_name": "1", "article_tags": "ds,ml,dl"},
                        {"article_name": "2", "article_tags": "ds,ml"}])
x_test = pd.DataFrame([{"article_name": "3", "article_tags": "ds,ml,py"}])

x_train_with_tags, x_test_with_tags = append_tags_to_frame(x_train, x_test, "article_tags", "tag_")

The result will be:

x_train_with_tags:

article_name tag_ds tag_ml tag_dl
1 1 1 1
2 1 1 0

x_test_with_tags:

article_name tag_ds tag_ml tag_dl
3 1 1 0

Extract Significant Terms from Subset

This method returns interesting or unusual occurrences of terms in a subset. It is based on the elasticsearch significant_text aggregation.

import pandas as pd
from ds_utils.strings import extract_significant_terms_from_subset

corpus = ['This is the first document.', 'This document is the second document.',
          'And this is the third one.', 'Is this the first document?']
data_frame = pd.DataFrame(corpus, columns=["content"])
# Let's differentiate between the last two documents from the full corpus
subset_data_frame = data_frame[data_frame.index > 1]
terms = extract_significant_terms_from_subset(data_frame, subset_data_frame,
                                              "content")

The output for terms will be the following table:

third one and this the is first document second
1.0 1.0 1.0 0.67 0.67 0.67 0.5 0.25 0.0

Unsupervised

Cluster Cardinality

Cluster cardinality is the number of examples per cluster. This method plots the number of points per cluster as a bar chart.

import pandas as pd
from matplotlib import pyplot as plt
from sklearn.cluster import KMeans
from ds_utils.unsupervised import plot_cluster_cardinality

data = pd.read_csv(path / to / dataset)
estimator = KMeans(n_clusters=8, random_state=42)
estimator.fit(data)

plot_cluster_cardinality(estimator.labels_)

plt.show()

Cluster Cardinality

Plot Cluster Magnitude

Cluster magnitude is the sum of distances from all examples to the centroid of the cluster. This method plots the Total Point-to-Centroid Distance per cluster as a bar chart.

import pandas as pd
from matplotlib import pyplot as plt
from sklearn.cluster import KMeans
from scipy.spatial.distance import euclidean
from ds_utils.unsupervised import plot_cluster_magnitude

data = pd.read_csv(path / to / dataset)
estimator = KMeans(n_clusters=8, random_state=42)
estimator.fit(data)

plot_cluster_magnitude(data, estimator.labels_, estimator.cluster_centers_, euclidean)

plt.show()

Plot Cluster Magnitude

Magnitude vs. Cardinality

Higher cluster cardinality tends to result in a higher cluster magnitude, which intuitively makes sense. Clusters are considered anomalous when cardinality doesn't correlate with magnitude relative to the other clusters. This method helps find anomalous clusters by plotting magnitude against cardinality as a scatter plot.

import pandas as pd
from matplotlib import pyplot as plt
from sklearn.cluster import KMeans
from scipy.spatial.distance import euclidean
from ds_utils.unsupervised import plot_magnitude_vs_cardinality

data = pd.read_csv(path / to / dataset)
estimator = KMeans(n_clusters=8, random_state=42)
estimator.fit(data)

plot_magnitude_vs_cardinality(data, estimator.labels_, estimator.cluster_centers_, euclidean)

plt.show()

Magnitude vs. Cardinality

Optimum Number of Clusters

K-means clustering requires you to decide the number of clusters k beforehand. This method runs the KMeans algorithm and increases the cluster number at each iteration. The total magnitude or sum of distances is used as the loss metric.

Note: Currently, this method only works with sklearn.cluster.KMeans.

import pandas as pd
from matplotlib import pyplot as plt
from scipy.spatial.distance import euclidean
from ds_utils.unsupervised import plot_loss_vs_cluster_number

data = pd.read_csv(path / to / dataset)

plot_loss_vs_cluster_number(data, 3, 20, euclidean)

plt.show()

Optimum Number of Clusters

XAI (Explainable AI)

Plot Feature Importance

This method plots feature importance as a bar chart, helping to visualize which features have the most significant impact on the model's decisions.

import pandas as pd
from matplotlib import pyplot as plt
from sklearn.tree import DecisionTreeClassifier
from ds_utils.xai import plot_features_importance

# Load the dataset
data = pd.read_csv(path / to / dataset)
target = data["target"]
features = data.columns.tolist()
features.remove("target")

# Train a decision tree classifier
clf = DecisionTreeClassifier(random_state=42)
clf.fit(data[features], target)

# Plot feature importance
plot_features_importance(features, clf.feature_importances_)

plt.show()

Plot Features Importance

This visualization helps in understanding which features are most influential in the model's decision-making process, providing valuable insights for feature selection and model interpretation.

Explore More

Excited about what you've seen so far? There's even more to discover! Dive deeper into each module to unlock the full potential of DataScienceUtils:

  • Metrics - Powerful methods for calculating and visualizing algorithm performance evaluation. Gain insights into how your models are performing.

  • Preprocess - Essential data preprocessing techniques to prepare your data for training. Improve your model's input for better results.

  • Strings - Efficient methods for manipulating and processing strings in dataframes. Handle text data with ease.

  • Unsupervised - Tools for calculating and visualizing the performance of unsupervised models. Understand your clustering and dimensionality reduction results better.

  • XAI - Methods to help explain model decisions, making your AI more interpretable and trustworthy.

Each module is designed to streamline your data science workflow, providing you with the tools you need to preprocess data, train models, evaluate performance, and interpret results. Check out the detailed documentation for each module to see how DataScienceUtils can enhance your projects!

Contributing

We're thrilled that you're interested in contributing to Data Science Utils! Your contributions help make this project better for everyone. Whether you're a seasoned developer or just getting started, there's a place for you here.

How to Contribute

  1. Find an area to contribute to: Check out our issues page for open tasks, or think of a feature you'd like to add.

  2. Fork the repository: Make your own copy of the project to work on.

  3. Create a branch: Make your changes in a new git branch.

  4. Make your changes: Add your improvements or fixes. We appreciate:

    • Bug reports and fixes
    • Feature requests and implementations
    • Documentation improvements
    • Performance optimizations
    • User experience enhancements
  5. Test your changes: Ensure your code works as expected and doesn't introduce new issues.

  6. Submit a pull request: Open a PR with a clear title and description of your changes.

Coding Guidelines

We follow the Python Software Foundation Code of Conduct and the Matplotlib Usage Guide. Please adhere to these guidelines in your contributions.

Getting Help

If you're new to open source or need any help, don't hesitate to ask questions in the issues section or reach out to the maintainers. We're here to help!

Why Contribute?

  • Improve your skills: Gain experience working on a real-world project.
  • Be part of a community: Connect with other developers and data scientists.
  • Make a difference: Your contributions will help others in their data science journey.
  • Get recognition: All contributors are acknowledged in our project.

Remember, no contribution is too small. Whether it's fixing a typo in documentation or adding a major feature, all contributions are valued and appreciated.

Thank you for helping make Data Science Utils better for everyone!

Installation Guide

Here are several ways to install the package:

1. Install from PyPI (Recommended)

The simplest way to install Data Science Utils and its dependencies is from PyPI using pip, Python's preferred package installer:

pip install data-science-utils

To upgrade Data Science Utils to the latest version, use:

pip install -U data-science-utils

2. Install from Source

If you prefer to install from source, you can clone the repository and install:

git clone https://github.com/idanmoradarthas/DataScienceUtils.git
cd DataScienceUtils
pip install .

Alternatively, you can install directly from GitHub using pip:

pip install git+https://github.com/idanmoradarthas/DataScienceUtils.git

3. Install using Anaconda

If you're using Anaconda, you can install using conda:

conda install idanmorad::data-science-utils

Note on Dependencies

Data Science Utils has several dependencies, including numpy, pandas, matplotlib, plotly and scikit-learn. These will be automatically installed when you install the package using the methods above.

Staying Updated

Data Science Utils is an active project that routinely publishes new releases with additional methods and improvements. We recommend periodically checking for updates to access the latest features and bug fixes.

If you encounter any issues during installation, please check our GitHub issues page or open a new issue for assistance.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

data_science_utils-1.8.0.tar.gz (33.6 kB view details)

Uploaded Source

Built Distribution

data_science_utils-1.8.0-py3-none-any.whl (24.3 kB view details)

Uploaded Python 3

File details

Details for the file data_science_utils-1.8.0.tar.gz.

File metadata

  • Download URL: data_science_utils-1.8.0.tar.gz
  • Upload date:
  • Size: 33.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.4

File hashes

Hashes for data_science_utils-1.8.0.tar.gz
Algorithm Hash digest
SHA256 b39ab8a040b893188fb1f73628ecb565129c5b542a49b52b095e3145a41288e5
MD5 4352ec8a8fdf4d8b8604478f8bf811f0
BLAKE2b-256 04a513fbd7ce30862d8285cfa08716cc7bc3b0c8d01fdee60d75f3e1226b3665

See more details on using hashes here.

File details

Details for the file data_science_utils-1.8.0-py3-none-any.whl.

File metadata

File hashes

Hashes for data_science_utils-1.8.0-py3-none-any.whl
Algorithm Hash digest
SHA256 6d6feee9d0e652c7e62dc07dbf2c32c2408cdab89d354ce2b7a1e5960db0d742
MD5 7bd1e3d61e21373677ba2d09478629f2
BLAKE2b-256 b7f7377193f37c0e7554eac67acc63093b6d0a83ce177e96ac39729e3d344505

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page