Skip to main content

Helper files/functions/classes for generic PySpark processes

Project description

toolbox-pyspark

github-release implementation version python-versions
os pypi-status pypi-format github-license pypi-downloads codecov-repo style
contributions
CI CD

Introduction

The purpose of this package is to provide some helper files/functions/classes for generic PySpark processes.

Key URLs

For reference, these URL's are used:

Type Source URL
Git Repo GitHub https://github.com/data-science-extensions/toolbox-pyspark
Python Package PyPI https://pypi.org/project/toolbox-pyspark
Package Docs Pages https://data-science-extensions.com/toolbox-pyspark

Installation

You can install and use this package multiple ways by using pip, pipenv, or poetry.

Using pip:

  1. In your terminal, run:

    python3 -m pip install --upgrade pip
    python3 -m pip install toolbox-pyspark
    
  2. Or, in your requirements.txt file, add:

    toolbox-pyspark
    

    Then run:

    python3 -m pip install --upgrade pip
    python3 -m pip install --requirement=requirements.txt
    

Using pipenv:

  1. Install using environment variables:

    In your Pipfile file, add:

    [[source]]
    url = "https://pypi.org/simple"
    verify_ssl = false
    name = "pypi"
    
    [packages]
    toolbox-pyspark = "*"
    

    Then run:

    python3 -m pip install pipenv
    python3 -m pipenv install --verbose --skip-lock --categories=root index=pypi toolbox-pyspark
    
  2. Or, in your requirements.txt file, add:

    toolbox-pyspark
    

    Then run:

    python3 -m run pipenv install --verbose --skip-lock --requirements=requirements.txt
    
  3. Or just run this:

    python3 -m pipenv install --verbose --skip-lock toolbox-pyspark
    

Using poetry:

  1. In your pyproject.toml file, add:

    [tool.poetry.dependencies]
    toolbox-pyspark = "*"
    

    Then run:

    poetry install
    
  2. Or just run this:

    poetry add toolbox-pyspark
    poetry install
    poetry sync
    

Contribution

Contribution is always welcome.

  1. First, either fork or branch the main repo.

  2. Clone your forked/branched repo.

  3. Build your environment:

    1. With pipenv on Windows:

      if (-not (Test-Path .venv)) {mkdir .venv}
      python -m pipenv install --requirements requirements.txt --requirements requirements-dev.txt --skip-lock
      python -m poetry run pre-commit install
      python -m poetry shell
      
    2. With pipenv on Linux:

      mkdir .venv
      python3 -m pipenv install --requirements requirements.txt --requirements requirements-dev.txt --skip-lock
      python3 -m poetry run pre-commit install
      python3 -m poetry shell
      
    3. With poetry on Windows:

      python -m pip install --upgrade pip
      python -m pip install poetry
      python -m poetry init
      python -m poetry add $(cat requirements/root.txt)
      python -m poetry add --group=dev $(cat requirements/dev.txt)
      python -m poetry add --group=test $(cat requirements/test.txt)
      python -m poetry add --group=docs $(cat requirements/docs.txt)
      python -m poetry install
      python -m poetry run pre-commit install
      python -m poetry shell
      
    4. With poetry on Linux:

      python3 -m pip install --upgrade pip
      python3 -m pip install poetry
      python3 -m poetry init
      python3 -m poetry add $(cat requirements/root.txt)
      python3 -m poetry add --group=dev $(cat requirements/dev.txt)
      python3 -m poetry add --group=test $(cat requirements/test.txt)
      python3 -m poetry add --group=docs $(cat requirements/docs.txt)
      python3 -m poetry install
      python3 -m poetry run pre-commit install
      python3 -m poetry shell
      
  4. Start contributing.

  5. When you're happy with the changes, raise a Pull Request to merge with the main branch again.

Build and Test

To ensure that the package is working as expected, please ensure that:

  1. You write your code as per PEP8 requirements.
  2. You write a UnitTest for each function/feature you include.
  3. The CodeCoverage is 100%.
  4. All UnitTests are passing.
  5. MyPy is passing 100%.

Testing

  • Run them all together

    poetry run make check
    
  • Or run them individually:

    • Black

      poetry run make check-black
      
    • PyTests:

      poetry run make ckeck-pytest
      
    • MyPy:

      poetry run make check-mypy
      

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

toolbox_pyspark-0.10.0.tar.gz (46.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

toolbox_pyspark-0.10.0-py3-none-any.whl (53.6 kB view details)

Uploaded Python 3

File details

Details for the file toolbox_pyspark-0.10.0.tar.gz.

File metadata

  • Download URL: toolbox_pyspark-0.10.0.tar.gz
  • Upload date:
  • Size: 46.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.5 CPython/3.13.1 Linux/6.5.0-1025-azure

File hashes

Hashes for toolbox_pyspark-0.10.0.tar.gz
Algorithm Hash digest
SHA256 9e3e33b4ce702780917b03336af8ad6c96c60b5a39ff00d7372d8f995a61bee2
MD5 71d365aaf3b7000255b926ccd49cf667
BLAKE2b-256 ae0e18fb546c03f5fbea2f683517f1999ef2b0702d60b5673a48c80e51065ffd

See more details on using hashes here.

File details

Details for the file toolbox_pyspark-0.10.0-py3-none-any.whl.

File metadata

  • Download URL: toolbox_pyspark-0.10.0-py3-none-any.whl
  • Upload date:
  • Size: 53.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.5 CPython/3.13.1 Linux/6.5.0-1025-azure

File hashes

Hashes for toolbox_pyspark-0.10.0-py3-none-any.whl
Algorithm Hash digest
SHA256 83a9b4a9a2a59e045acf2fd899047276273be06e30da37cd77e0b086c6bfea9a
MD5 989c722f71cc285c60d1f38d2674f7e1
BLAKE2b-256 9b25aed7a6c57d8e8ca0a1fb0203846938c7aa03291747b0adfd545925120284

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page