spinesUtils is a user-friendly toolkit for python development.
Project description
spinesUtils
Accelerate your Python development workflow
Overview
spinesUtils is a powerful library that provides ready-to-use features and utilities for Python development to shorten the project development cycle. Our goal is to help developers focus on solving their core problems instead of reimplementing common functionality.
Features
- Logging functionality - Simplified logging without handler conflicts
- Type checking and parameter validation - Robust validation decorators
- CSV file reading acceleration - Performance-optimized data loading
- Imbalanced data classifiers - Specialized ML tools for imbalanced datasets
- Pandas DataFrame data compression - Memory optimization for large datasets
- DataFrame insight tools - Quick data analysis and visualization
- Large data train-test splitting - Efficient data partitioning for ML pipelines
- Intuitive timer - Simple timing and benchmarking tools
This library is currently undergoing rapid iteration. If you encounter any issues with its functionalities, feel free to raise an issue.
Installation
You can install spinesUtils from PyPI:
pip install spinesUtils
Usage Examples
Logger
The Logger class provides convenient logging without worrying about handler conflicts with the native Python logging module.
# load spinesUtils module
from spinesUtils.logging import Logger
# create a logger instance, with name "MyLogger", and no file handler, the default level is "INFO"
# You can specify a file path `fp` during instantiation. If not specified, logs will not be written to a file.
logger = Logger(name="MyLogger", fp=None, level="DEBUG")
logger.log("This is an info log emitted by the log function.", level='INFO')
logger.debug("This is an debug message")
logger.info("This is an info message.")
logger.warning("This is an warning message.")
logger.error("This is an error message.")
logger.critical("This is an critical message.")
Type Checking and Parameter Validation
Ensure your functions receive the correct input types and values:
from spinesUtils.asserts import *
# Check parameter type
@ParameterTypeAssert({
'a': (int, float),
'b': (int, float)
})
def add(a, b):
return a + b
# Check parameter value
@ParameterValuesAssert({
'a': lambda x: x > 0,
'b': lambda x: x > 0
})
def divide(a, b):
return a / b
# Generate function kwargs
params = generate_function_kwargs(add, a=1, b=2)
CSV Reading Acceleration
Read large CSV files efficiently:
from spinesUtils import read_csv
df = read_csv(
fp='/path/to/your/file.csv',
sep=',', # equal to pandas read_csv.sep
turbo_method='polars', # use turbo_method to speed up load time
chunk_size=None, # it can be integer if you want to use pandas backend
transform2low_mem=True, # compresses file to save memory
verbose=False
)
Classifiers for Imbalanced Data
Handle imbalanced datasets effectively:
from spinesUtils.models import MultiClassBalanceClassifier
from sklearn.ensemble import RandomForestClassifier
classifier = MultiClassBalanceClassifier(
base_estimator=RandomForestClassifier(n_estimators=100),
n_classes=3,
random_state=0,
verbose=0
)
# Fit and predict as you would with any scikit-learn estimator
classifier.fit(X_train, y_train)
y_pred = classifier.predict(X_test)
DataFrame Data Compression
Optimize memory usage for large DataFrames:
from spinesUtils import transform_dtypes_low_mem
# Compress a single DataFrame
transform_dtypes_low_mem(df, verbose=True, inplace=True)
# Batch compress multiple DataFrames
from spinesUtils import transform_batch_dtypes_low_mem
transform_batch_dtypes_low_mem([df1, df2, df3, df4], verbose=True, inplace=True)
DataFrame Insight Tools
Quickly analyze your data:
from spinesUtils import df_preview, classify_samples_dist
# Get comprehensive DataFrame insights
df_insight = df_preview(df)
Data Splitting Utilities
Efficiently split large datasets:
from spinesUtils import train_test_split_bigdata, train_test_split_bigdata_df
from spinesUtils.feature_tools import get_x_cols
# Return numpy arrays
X_train, X_valid, X_test, y_train, y_valid, y_test = train_test_split_bigdata(
df=df,
x_cols=get_x_cols(df, y_col='target_column'),
y_col='target_column',
shuffle=True,
return_valid=True,
train_size=0.8,
valid_size=0.5
)
# Return pandas DataFrames
train_df, valid_df, test_df = train_test_split_bigdata_df(
df=df,
x_cols=get_x_cols(df, y_col='target_column'),
y_col='target_column',
shuffle=True,
return_valid=True,
train_size=0.8,
valid_size=0.5
)
Timer Utility
Time your code execution simply:
from spinesUtils.timer import Timer
# As a context manager
with Timer().session() as t:
# Your code here
t.sleep(1)
print(f"Step 1 time: {t.last_timestamp_diff():.2f}s")
# Mark a middle point
t.middle_point()
# More code
t.sleep(2)
print(f"Step 2 time: {t.last_timestamp_diff():.2f}s")
print(f"Total time: {t.total_elapsed_time():.2f}s")
# Or use it manually
timer = Timer()
timer.start()
# Your code here
timer.end()
print(f"Elapsed: {timer.total_elapsed_time():.2f}s")
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file spinesutils-0.5.0.tar.gz.
File metadata
- Download URL: spinesutils-0.5.0.tar.gz
- Upload date:
- Size: 52.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.11.10
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
fa78ae0745ac6b15fc000975f389749228b26b7748d8429e14d0fc4e1debdac7
|
|
| MD5 |
b2c09fa9f94b4af570e53dbf70d1daf1
|
|
| BLAKE2b-256 |
f1cd2b509391abc0d7be54bff6853dcf6812e9c13937d331fc9d1ed70f404446
|
File details
Details for the file spinesutils-0.5.0-py3-none-any.whl.
File metadata
- Download URL: spinesutils-0.5.0-py3-none-any.whl
- Upload date:
- Size: 45.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.1.0 CPython/3.11.10
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
bff954ec7b4da8cb93e9a4b6db40b11f67cfd8b46b37b37667977cd93240162d
|
|
| MD5 |
b6c7e0c375100378a158382ec077f344
|
|
| BLAKE2b-256 |
4d7989e9ba5c63181dd47775aa94f9e65e334e532cc95c5dcd718a915e7bccff
|