Skip to main content

No project description provided

Project description

Feature Pipeline

Check out this Medium article for more details about this module.

Create Environment File

~/energy-forecasting $ cp .env.default .env

The command cp .env.default .env is used to create a copy of the .env.default file and name it .env. In many projects, the .env file is used to store environment variables that the application needs to run. The .env.default file is usually a template that includes all the environment variables that the application expects, but with default values. By copying it to .env, you can customize these values for your own environment.

Set Up the ML_PIPELINE_ROOT_DIR Variable

~/energy-forecasting $ export ML_PIPELINE_ROOT_DIR=$(pwd)

The command export ML_PIPELINE_ROOT_DIR=$(pwd) is setting the value of the ML_PIPELINE_ROOT_DIR environment variable to the current directory. In this context, $(pwd) is a command substitution that gets replaced with the output of the pwd command, which prints the path of the current directory. The export command then makes this variable available to child processes of the current shell.

In essence, ML_PIPELINE_ROOT_DIR is an environment variable that is set to the path of the current directory. This can be useful for scripts or programs that need to reference the root directory of the ML pipeline, as they can simply refer to ML_PIPELINE_ROOT_DIR instead of needing to know the exact path.

Install for Development

Create virtual environment:

~/energy-forecasting                  $ cd feature-pipeline && rm poetry.lock
~/energy-forecasting/feature-pipeline $ bash ../scripts/devops/virtual_environment/poetry_install.sh
~/energy-forecasting/feature-pipeline $ source .venv/bin/activate
  1. We first navigate to the feature-pipeline directory and remove the poetry.lock file. This step is essential if we intend to add new dependencies to the pyproject.toml file, as it ensures that Poetry accurately resolves and installs the latest compatible versions of all dependencies.
  2. We then execute the poetry_install.sh script. This script is responsible for creating the virtual environment and installing the project dependencies. Importantly, it also includes steps to resolve potential issues related to the macOS arm64 architecture.
  3. Finally, we activate the virtual environment. This step provides an isolated workspace for our project, preventing conflicts between the project's dependencies and those installed globally on the system.

Check the Set Up Additional Tools and Usage sections to see how to set up the additional tools and credentials you need to run this project.

Usage for Development

To start the ETL pipeline run:

~/energy-forecasting/feature-pipeline $ python -m feature_pipeline.pipeline

To create a new feature view run:

~/energy-forecasting/feature-pipeline $ python -m feature_pipeline.feature_view

NOTE: Be careful to set the ML_PIPELINE_ROOT_DIR variable as explained in this section.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

g_feature_pipeline-0.2.0.tar.gz (9.2 kB view details)

Uploaded Source

Built Distribution

g_feature_pipeline-0.2.0-py3-none-any.whl (11.3 kB view details)

Uploaded Python 3

File details

Details for the file g_feature_pipeline-0.2.0.tar.gz.

File metadata

  • Download URL: g_feature_pipeline-0.2.0.tar.gz
  • Upload date:
  • Size: 9.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.5.1 CPython/3.10.9 Darwin/22.3.0

File hashes

Hashes for g_feature_pipeline-0.2.0.tar.gz
Algorithm Hash digest
SHA256 c885f7ce6b511e1174b487c2bf479f376e17d3f9f6b420ca5b216ebeb989b0a2
MD5 899528b9e28e144895a40679d6a7bdde
BLAKE2b-256 e19f30592afc6b38471f68216fb542817dd45bb28e26a20b74caf23def969087

See more details on using hashes here.

File details

Details for the file g_feature_pipeline-0.2.0-py3-none-any.whl.

File metadata

File hashes

Hashes for g_feature_pipeline-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 1d38574ea4119aeda227fe1d850b812c36168c9d276fb5745e7c899bcc0eb586
MD5 08c4aef8aefff298bedaa7070f5208d6
BLAKE2b-256 6baf7c34a260f6f39c917afe48a88d5dde53721c0103c26afb1ac5dfa6fa7f40

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page