Skip to main content

Framework for machine and deep learning, with regression, classification and time series analysis

Project description

🐸

Welcome to LeCrapaud

An all-in-one machine learning framework

🚀 Introduction

LeCrapaud is a high-level Python library for end-to-end machine learning workflows on tabular data, with a focus on financial and stock datasets. It provides a simple API to handle feature engineering, model selection, training, and prediction, all in a reproducible and modular way.

✨ Key Features

  • 🧩 Modular pipeline: Feature engineering, preprocessing, selection, and modeling as independent steps
  • 🤖 Automated model selection and hyperparameter optimization
  • 📊 Easy integration with pandas DataFrames
  • 🔬 Supports both regression and classification tasks
  • 🛠️ Simple API for both full pipeline and step-by-step usage
  • 📦 Ready for production and research workflows

⚡ Quick Start

  1. Create environment
$ pip install virtualenv
$ python -m venv .venv
$ source .venv/bin/activate
  1. Install dependencies
$ make install
  1. Deactivate virtualenv (if needed)
$ deactivate

🛠️ How it works

This package provides a high-level API to manage experiments for feature engineering, model selection, and prediction on tabular data (e.g. stock data).

Typical workflow

from lecrapaud.api import LeCrapaud

# 1. Create the main app
app = LeCrapaud()

# 2. Define your experiment context (see your notebook or api.py for all options)
context = {
    "data": your_dataframe,
    "columns_drop": [...],
    "columns_date": [...],
    # ... other config options
}

# 3. Create an experiment
experiment = app.create_experiment(**context)

# 4. Run the full training pipeline
experiment.train(your_dataframe)

# 5. Make predictions on new data
predictions = experiment.predict(new_data)

Modular usage

You can also use each step independently:

data_eng = experiment.feature_engineering(data)
train, val, test = experiment.preprocess_feature(data_eng)
features = experiment.feature_selection(train)
std_data, reshaped_data = experiment.preprocess_model(train, val, test)
experiment.model_selection(std_data, reshaped_data)

🤝 Contributing

Reminders for Github usage

  1. Creating Github repository
$ brew install gh
$ gh auth login
$ gh repo create
  1. Initializing git and first commit to distant repository
$ git init
$ git add .
$ git commit -m 'first commit'
$ git remote add origin <YOUR_REPO_URL>
$ git push -u origin master
  1. Use conventional commits
    https://www.conventionalcommits.org/en/v1.0.0/#summary

Pierre Gallet © 2024

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lecrapaud-0.2.0.tar.gz (85.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lecrapaud-0.2.0-py3-none-any.whl (111.9 kB view details)

Uploaded Python 3

File details

Details for the file lecrapaud-0.2.0.tar.gz.

File metadata

  • Download URL: lecrapaud-0.2.0.tar.gz
  • Upload date:
  • Size: 85.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.10

File hashes

Hashes for lecrapaud-0.2.0.tar.gz
Algorithm Hash digest
SHA256 81eef69e0139f82ee4bb86becdf9532ebc23e541ed5d76a3f757b749e1df1161
MD5 0da8699885d7059cec6cb3010d52c339
BLAKE2b-256 6d94bd3f4560c40f346095cb8c97b8fc700bf3838c17d292968f4cc1cd261293

See more details on using hashes here.

File details

Details for the file lecrapaud-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: lecrapaud-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 111.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.10

File hashes

Hashes for lecrapaud-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 def099b9277a9c5e93574eb7991186f91a49d145870d647205d5980e74b9e045
MD5 bf5760473c47a1b17a9b79ceb980ea18
BLAKE2b-256 5507834b8c975b1d297da73f5d44fab2ad01eb928e6a4873f13604d36a2403b2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page