Python GUI tool to collect metadata for DSP projects.
Project description
DSP-METADATA-GUI Metadata Module
The dsp-metadata-gui
is a GUI application written in Python for collecting project specific metadata and turn it into RDF.
As part of the dsp-tools
, its aim is to enable researchers and project managers who deposit research data on the DaSCH Service Platform (DSP), to add metadata about the project and datasets to the DSP repository. By providing metadata, the project will be searchable on the platform, which is an integral part of the FAIR principles.
The metadata follows the schema defined by the dsp-ontologies.
Install and run
The module provides a command line entry point to run the GUI. The only requirement is Python 3 and PIP.
The application has only been tested on Python 3.9, but it might work on older versions too.
Note: There is a number of known potential issues. See the troubleshoot
section here.
Installation via pip
To install the application, run:
pip install dsp-metadata-gui
Or respectively:
pip3 install dsp-metadata-gui
Afterwards, the program can be started by running the command dsp-metadata
in your terminal of choice.
Installation from source
Clone this repo and run make make-and-run
. If you don't use GNU Make, run the commands specified in the Makefile
manually.
This will package the application, install it to your python environment and run the application.
Usage
Collecting Metadata
The application is divided into two windows:
-
The main window lets you organize a list of projects, for which you can collect metadata. Several actions can be performed with projects, e.g. editing or exporting the project.
-
When editing a project, in the project window, the actual metadata can be added, modified and saved.
To add a project, you will need the project short code, which is assigned to you by the DaSCH Client Services.
A project is always associated with a folder on your local machine. If any files should be included with the metadata import, these files must be within that folder.
Once all metadata are added and valid, and the overall RDF graph of the metadata set validates against the ontology, the project can be exported for upload to the DSP.
All data is locally stored in the file ~/DaSCH/config/repos.data
. for more detail, see here.
Conversion to V2
The metadata generated by the application conforms to the first version of the data model for metadata.
This corresponds to the data that can currently be viewed in the DaSCH Metadata Browser.
The initial data model will eventually be replaced by the model V2 which introduces major improvements.
Metadata V2 will eventually be collected directly in the web interface rather than in this python application.
In the mean time until the web interface for editing metadata is implemented, this application provides a script to automatically convert V1 .ttl
files into V2 .json
files.
NB: The conversion can not be fully automated, as the model V2 is more rich in information than V1.
For convenience, the conversion adds the stringXX
wherever the output can not be determined with sufficient confidence. Please check those instances manually.
The conversion also does some "guessing" work, as e.g. the language of literal values or the display text for URLs. If the output can be determined with a sufficient level of confidence, the conversion will not addXX
. However it is still advisable to check the entirety of the output for potential errors.
V2 JSON metadata can again be converted to V2 RDF metadata, using another script. This should not require any additional data cleaning.
The most important changes from V1 to V2 include the following additions:
-
Support for multi-language literals
-
howToCite
on project level -
country
property for addresses -
Creation and modification timestamps
-
JSON schema validation
Development
Development Environment
Pipenv
Use pipenv
for a seamless development experience.
In order to have both dependencies and dev-dependencies installed from the Pipfile
, set up the virtual environment running
pipenv instal --dev
pipenv
will manage dependencies as well as a virtual environment. To install packages, use
pipenv install <package-name>
To create requirements.txt
, run
pipenv lock -r > requirements.txt
To bring setup.py
up to date, run
pipenv run pipenv-setup sync
GNU Make
GNU Make
is used to automatize most tasks.
Run make help
for info on the available targets.
Note: All make targets - except make run
- should be run from within the pipenv
shell:
Either by running
pipenv run make <target-name>
or by opening a virtual pipenv shell by running
pipenv shell
make <target-name>
...
exit
Documentation
The documentation is created using mkdocs
and mkdocstrings
with markdown_include.include
. To create the documentation, make sure to install all of these, using pip.
To serve the documentation locally, run make doc
. To deploy the documentation to github pages, run make deploy-doc
.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for dsp_metadata_gui-1.2.0-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5360482235c9605b21d49a40bda7b05a6d1376aeb0f08a0453b42142547ff9f4 |
|
MD5 | a5c3980ff14e0a6c7169b7e53b7beb0a |
|
BLAKE2b-256 | 5cc304ad057b0cb9096b2a95ce9f45fc43c7da1043b0143e7bce57f5bc379181 |