Skip to main content

A Python library and tools for the DaSCH Service Platform

Project description

PyPI version

DSP-TOOLS - DaSCH Service Platform Tools

dsp-tools is a command line tool that helps you to interact with the DaSCH Service Platform API. Consult the full documentation here.

Information for developers

There is a Makefile for all the following tasks (and more). Type make to print the available targets.

For a quick start, use:

pip install pipenv
pipenv install --dev
pipenv run make install

This creates a pipenv-environment, installs all dependencies, and installs dsp-tools from source.

If you prefer getting around pipenv, use instead:

make install-requirements
make install

Git submodules

This repository embeds https://github.com/dasch-swiss/0123-import-scripts as a Git submodule in knora/dsplib/import_scripts. That means that knora/dsplib/import_scripts has no contents, but only a reference to a certain commit in the main branch of 0123-import-scripts. When you clone dsp-tools from GitHub as usual, knora/dsplib/import_scripts will be empty.

Passively using the contents of the submodule

If you don't have a clone of dsp-tools yet, clone it with

git clone --recurse-submodules https://github.com/dasch-swiss/dsp-tools.git

After cloning it that way, and after some time has passed, you might want to get the latest changes from GitHub:

cd dsp-tools
git pull --recurse-submodules

These two commands take care of the submodule, so that its contents are cloned/pulled as well.

In case you have an old clone of dsp-tools, without the submodule, and you want to update it, you have to proceed differently:

cd dsp-tools
git pull
git submodule update --init --recursive

Some notes:

  • git clone --recurse-submodules <repo> is shorthand for git clone <repo>; cd <repo>; git submodule update --init --recursive
  • git pull --recurse-submodules is shorthand for git pull; git submodule update --init --recursive
  • --init is necessary if you don't have the submodule knora/dsplib/import_scripts yet. In all successive calls, when the submodule is already on your machine, the flag --init can be omitted.
  • --recursive is optional, in case there would be more than one (nested) submodules inside dsp-tools.
  • Since Git 2.15, you can tell Git to use --recurse-submodules for all commands that support it (except clone), with git config submodule.recurse true.
  • These explanations rely on the official Git Submodules documentation

Actively working with the contents of the submodule

After retrieving the contents of a submodule as described in the paragraph above, it is in "detached HEAD" state. Before committing to it, the main branch needs to be checked out. The order how to proceed is the following:

cd knora/dsplib/import_scripts
git checkout main                     # check out main branch of 0123-import-scripts
# (modify contents of submodule)
git add .
git commit -m "modify submodule"
git push origin main                  # push to origin of 0123-import-scripts
cd ../../..
git add knora/dsplib/import_scripts
git commit -m "modify submodule"
git push origin feature-branch        # push to origin of dsp-tools

When switching between branches, there are two options:

  1. By default (submodule.recurse is false AND branches are switched with git checkout <branch>), the contents of submodules will not be updated.
  2. If submodule.recurse has been set to true, OR if branches are switched with git checkout <branch> --recurse-submodules, the contents of submodules will be updated according to the commit recorded in the superproject. If local modifications in a submodule would be overwritten, the checkout will fail.

To quickly switch between branches when you have uncommitted work in the submodule, the first option might be preferable. After merging a Pull Request and switching back to the main branch, the second option might be more suitable.
Read more about the checkout options in the official documentation

Pipenv

We use pipenv for our dependency management. There are two ways to get started:

  • pipenv install --dev installs all dependencies, while giving them the opportunity to update themselves
  • pipenv install --ignore-pipfile is used to get a deterministic build in production

This works because there are two files defining the dependencies:

  • Pipfile replaces requirements.txt, but lists only the core dependencies, ordered in two sections:
    • [packages] lists the dependencies used to run the software.
    • [dev-packages] lists additional dependencies used for tests and deployment.
  • Pipfile.lock enables deterministic builds, by exactly pinning the version of all (sub-) dependencies. This is done automatically, you must not edit Pipfile.lock.

The diverse requirements.txt files in this repo are only present for backwards compatibility and for GitHub CI.

If you want to install a new package, install it with pipenv install package. This

  • installs the package (incl. sub-dependencies) in your virtual environment
  • adds the package to the section [packages] of Pipfile. By default, no versions are pinned
  • adds the pinned versions of package and all sub-dependencies to Pipfile.lock

If a package is only needed for development, please install it with pipenv install package --dev, so it gets added to the [dev-packages] section of Pipfile.

For security reasons, the maintainer regularly executes

  • pipenv check to get informed about vulnerabilities
  • pipenv update to update Pipfile.lock with the latest version of every package
  • make freeze-requirements to update the requirement files and setup.py

make freeze-requirements must also be executed after adding a new dependency. If you prefer working without pipenv, you can freeze your requirements with pip3 freeze > requirements.txt and update setup.py manually.

Pipenv setup in PyCharm

  • Go to Add Interpreter > Pipenv Environment
  • Base Interpreter: PyCarm auto-detects one of your system-wide installed Pythons as base interpreter.
  • Pipenv executable: auto-detected
  • After hitting OK, PyCharm creates a new pipenv environment and installs the dependencies from Pipfile

If you already initialized a pipenv-environment via command line, you can add its interpreter in PyCharm, but this will create the pipenv-environment again.

Testing

Please note that testing requires launching the complete DSP API stack which is based on docker images. Therefore, we recommend installing the docker desktop client.
To run the complete test suite:

make test

Code style

When contributing to the project please make sure you use the same code style rules as we do. We use autopep8 and mypy. The configuration is defined in pyproject.toml in the root directory of the project.

You can use the configuration with autopep8 --global-config pyproject.toml [file path] and mypy --config-file pyproject.toml [file path].

If you are using PyCharm we recommend installing autopep8 as external tool. You can then use it with right-click on the file > External Tools > autopep8 to reformat files in-place. Due to compatibility issues with VSCode, the argument --in-place=true can not be declared in the pyproject.toml and needs to be passed to the external tool in the PyCharm settings.
mypy is available as plugin.

In VSCode, both mypy and autopep8 can be set up as default linter and formatter through the python extension.

For formatting Markdown files (*.md) we use the default styling configuration provided by PyCharm.

Publishing

Publishing is automated with GitHub Actions and should not be done manually. Please follow the Pull Request Guidelines. If done correctly, when merging a pull request into main, the release-please action will create or update a pull request for a release. This pull request will follow semantic versioning and update the change log. Once all desired features are merged, the release can be executed by merging this release pull request into main. This will trigger actions that create a release on GitHub, on PyPI and the docs.

Please ensure you have only one pull request per feature.

Publishing manually

Publishing is automated with GitHub Actions and should not be done manually. If you still need to do it, follow the steps below.

Generate the distribution package. Make sure you have the latest versions of setuptools and wheel installed:

python3 -m pip install --upgrade pip setuptools wheel
python3 setup.py sdist bdist_wheel

You can install the package locally from the dist:

python3 -m pip ./dist/some_name.whl

Upload package works also with make:

make dist
make upload

For local development:

python3 setup.py develop

Contributing to the documentation

The documentation is a collection of markdown files in the docs folder.
After updates of the files, build and check the result with the following command:

make docs-serve 

The documentation is published on https://docs.dasch.swiss/latest/DSP-TOOLS. During the centralized release process of all components of the DSP software stack, the docs of dsp-tools get built from the main branch to https://docs.dasch.swiss.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

dsp-tools-1.22.0.tar.gz (125.7 kB view details)

Uploaded Source

Built Distribution

dsp_tools-1.22.0-py3-none-any.whl (148.3 kB view details)

Uploaded Python 3

File details

Details for the file dsp-tools-1.22.0.tar.gz.

File metadata

  • Download URL: dsp-tools-1.22.0.tar.gz
  • Upload date:
  • Size: 125.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.16

File hashes

Hashes for dsp-tools-1.22.0.tar.gz
Algorithm Hash digest
SHA256 52f9091ce2b7f99f8029a9f63870f1f1c28a0072e1d5fe1949d28eaa8d79efac
MD5 50dfdd4a92515dced853d6ef0a8d7400
BLAKE2b-256 461fde804733b17776712eed35fde4315fe2b046b1ef780d69638d1e56abfeda

See more details on using hashes here.

File details

Details for the file dsp_tools-1.22.0-py3-none-any.whl.

File metadata

  • Download URL: dsp_tools-1.22.0-py3-none-any.whl
  • Upload date:
  • Size: 148.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.16

File hashes

Hashes for dsp_tools-1.22.0-py3-none-any.whl
Algorithm Hash digest
SHA256 b56fbef1170fb8cef61620adbd16006b19c546ff8bcf59eb04cc3bb80cc643fb
MD5 c09d8fe5a60fab2e7c77878c1820bfa4
BLAKE2b-256 efbb18b12eae2b9c9a1b4ebae4292613464fa28710ed85e3fff6cbcf543bc34d

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page