A Python library and tools for the DaSCH Service Platform
Project description
DSP-TOOLS - DaSCH Service Platform Tools
dsp-tools is a command line tool that helps you to interact with the DaSCH Service Platform API. Consult the full documentation here.
Information for developers
There is a Makefile
for all the following tasks (and more). Type make
to print the available targets.
For a quick start, use:
pip install pipenv
pipenv install --dev
pipenv run make install
This creates a pipenv-environment, installs all dependencies, and installs dsp-tools
from source.
If you prefer getting around pipenv, use instead:
make install-requirements
make install
User data in the folder .dsp-tools
DSP-TOOLS saves user data in the user's home directory, in the folder .dsp-tools
. Here is an overview of its
structure:
folder | command using it | description |
---|---|---|
xmluploads | xmlupload |
saves id2iri mappings and error reports |
docker | stack-up |
files necessary to startup Docker containers |
The start-stack
command
This command starts Docker containers of DSP-API and DSP-APP, in the version that is running on https://admin.dasch. swiss. In addition to the containers, a number of files from the DSP-API GitHub repository is necessary. The version of the docker images and these files must be the same. The version is hardcoded at the following places in the code:
knora/dsplib/docker/docker-compose.yml
: The 4 variablesservices/{app,db,sipi,api}/image
must point to the DockerHub image of the last deployed versionknora/dsplib/utils/stack_handling.py
: The variablecommit_of_used_api_version
must be the commit hash of DSP-API of the version that is running on https://admin.dasch.swiss.
Git submodules
This repository embeds https://github.com/dasch-swiss/0123-import-scripts
as a Git submodule in knora/dsplib/import_scripts
. That means that knora/dsplib/import_scripts
has no contents, but
only a reference to a certain commit in the main branch of 0123-import-scripts
. When you clone dsp-tools
from GitHub
as usual, knora/dsplib/import_scripts
will be empty.
Passively using the contents of the submodule
If you don't have a clone of dsp-tools yet, clone it with
git clone --recurse-submodules https://github.com/dasch-swiss/dsp-tools.git
After cloning it that way, and after some time has passed, you might want to get the latest changes from GitHub:
cd dsp-tools
git pull --recurse-submodules
These two commands take care of the submodule, so that its contents are cloned/pulled as well.
In case you have an old clone of dsp-tools, without the submodule, and you want to update it, you have to proceed differently:
cd dsp-tools
git pull
git submodule update --init --recursive
Some notes:
git clone --recurse-submodules <repo>
is shorthand forgit clone <repo>; cd <repo>; git submodule update --init --recursive
git pull --recurse-submodules
is shorthand forgit pull; git submodule update --init --recursive
--init
is necessary if you don't have the submoduleknora/dsplib/import_scripts
yet. In all successive calls, when the submodule is already on your machine, the flag--init
can be omitted.--recursive
is optional, in case there would be more than one (nested) submodules inside dsp-tools.- Since Git 2.15, you can tell Git to use
--recurse-submodules
for all commands that support it (exceptclone
), withgit config submodule.recurse true
. - These explanations rely on the official Git Submodules documentation
Actively working with the contents of the submodule
After retrieving the contents of a submodule as described in the paragraph above, it is in "detached HEAD" state. Before
committing to it, the main
branch needs to be checked out. The order how to proceed is the following:
cd knora/dsplib/import_scripts
git checkout main # check out main branch of 0123-import-scripts
# (modify contents of submodule)
git add .
git commit -m "modify submodule"
git push origin main # push to origin of 0123-import-scripts
cd ../../..
git add knora/dsplib/import_scripts
git commit -m "modify submodule"
git push origin feature-branch # push to origin of dsp-tools
When switching between branches, there are two options:
- By default (
submodule.recurse
is false AND branches are switched withgit checkout <branch>
), the contents of submodules will not be updated. - If
submodule.recurse
has been set to true, OR if branches are switched withgit checkout <branch> --recurse-submodules
, the contents of submodules will be updated according to the commit recorded in the superproject. If local modifications in a submodule would be overwritten, the checkout will fail.
To quickly switch between branches when you have
uncommitted work in the submodule, the first option might be preferable. After merging a Pull Request and switching
back to the main branch, the second option might be more suitable.
Read more about the checkout options in the official documentation
Pipenv
We use pipenv for our dependency management. There are two ways to get started:
pipenv install --dev
installs all dependencies, while giving them the opportunity to update themselvespipenv install --ignore-pipfile
is used to get a deterministic build in production
This works because there are two files defining the dependencies:
Pipfile
replacesrequirements.txt
, but lists only the core dependencies, ordered in two sections:[packages]
lists the dependencies used to run the software.[dev-packages]
lists additional dependencies used for tests and deployment.
Pipfile.lock
enables deterministic builds, by exactly pinning the version of all (sub-) dependencies. This is done automatically, you must not editPipfile.lock
.
The diverse requirements.txt
files in this repo are only present for backwards compatibility
and for GitHub CI.
If you want to install a new package, install it with pipenv install package
. This
- installs the package (incl. sub-dependencies) in your virtual environment
- adds the package to the section
[packages]
ofPipfile
. By default, no versions are pinned - adds the pinned versions of package and all sub-dependencies to
Pipfile.lock
If a package is only needed for development, please install it with pipenv install package --dev
,
so it gets added to the [dev-packages]
section of Pipfile
.
For security reasons, the maintainer regularly executes
pipenv check
to get informed about vulnerabilitiespipenv update
to updatePipfile.lock
with the latest version of every packagemake freeze-requirements
to update the requirement files andsetup.py
make freeze-requirements
must also be executed after adding a new dependency. If you prefer working
without pipenv, you can freeze your requirements with pip3 freeze > requirements.txt
and update setup.py
manually.
Pipenv setup in PyCharm
- Go to Add Interpreter > Pipenv Environment
- Base Interpreter: PyCarm auto-detects one of your system-wide installed Pythons as base interpreter.
- Pipenv executable: auto-detected
- After hitting OK, PyCharm creates a new pipenv environment and installs the dependencies from
Pipfile
If you already initialized a pipenv-environment via command line, you can add its interpreter in PyCharm, but this will create the pipenv-environment again.
Testing
Please note that testing requires launching the complete DSP API stack which is based on docker images.
Therefore, we recommend installing the docker desktop client.
To run the complete test suite:
make test
Code style
When contributing to the project please make sure you use the same code style rules as we do. We use
autopep8 and mypy. The
configuration is defined in pyproject.toml
in the root directory of the project.
You can use the configuration with autopep8 --global-config pyproject.toml [file path]
and
mypy --config-file pyproject.toml [file path]
.
If you are using PyCharm we recommend installing autopep8 as external tool. You can then use it with
right-click on the file > External Tools
> autopep8
to reformat files in-place. Due to compatibility
issues with VSCode, the argument --in-place=true
can not be declared in the pyproject.toml
and
needs to be passed to the external tool in the PyCharm settings.
mypy is available as plugin.
In VSCode, both mypy and autopep8 can be set up as default linter and formatter through the python extension.
For formatting Markdown files (*.md) we use the default styling configuration provided by PyCharm.
Publishing
Publishing is automated with GitHub Actions and should not be done manually. Please follow the
Pull Request Guidelines. If done
correctly, when merging a pull request into main
, the release-please
action will create or update a pull request for
a release. This pull request will follow semantic versioning and update the change log. Once all desired features are
merged, the release can be executed by merging this release pull request into main
. This will trigger actions that
create a release on GitHub, on PyPI and the docs.
Please ensure you have only one pull request per feature.
Publishing manually
Publishing is automated with GitHub Actions and should not be done manually. If you still need to do it, follow the steps below.
Generate the distribution package. Make sure you have the latest versions of setuptools
and wheel
installed:
python3 -m pip install --upgrade pip setuptools wheel
python3 setup.py sdist bdist_wheel
You can install the package locally from the dist:
python3 -m pip ./dist/some_name.whl
Upload package works also with make
:
make dist
make upload
For local development:
python3 setup.py develop
Contributing to the documentation
The documentation is a collection of markdown files in the docs
folder.
After updates of the files, build and check the result with the following command:
make docs-serve
The documentation is published on https://docs.dasch.swiss/latest/DSP-TOOLS. During the centralized release process of all components of the DSP software stack, the docs of dsp-tools get built from the main branch to https://docs.dasch.swiss.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file dsp-tools-1.22.2.tar.gz
.
File metadata
- Download URL: dsp-tools-1.22.2.tar.gz
- Upload date:
- Size: 126.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.16
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 64622e7c1b9be3b6dcf40556949e6d1686c713b3d3b1f419326d0c61cc16e7c0 |
|
MD5 | 2b9faaf7434f82b96ab182019921a529 |
|
BLAKE2b-256 | 54127199dfba54d6fc6656158d4530e1982e21d47ae71fdb77be9bad8cb059b2 |
File details
Details for the file dsp_tools-1.22.2-py3-none-any.whl
.
File metadata
- Download URL: dsp_tools-1.22.2-py3-none-any.whl
- Upload date:
- Size: 148.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.16
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 104ddcf2ae8258a852f231c3ddda920990c3a43312c5a57761ccc5c9e952118e |
|
MD5 | 48a7035bd55265ba36a1a4718d563ef1 |
|
BLAKE2b-256 | 4fbc51ef7e5ec1daf39437b9ab3080ca66aafce11726f674533e6e206c3ac4f5 |