Skip to main content

A tool for building import sheets from UTK legacy systems

Project description

UTK Exodus :flight_departure:

About

This application is a complete rewrite of the code used to migrate UTK content from Islandora 7 to Hyku.

Unlike the previous code, this aims to be more flexible, easier to understand, and easier to use as a whole.

Installing

To install for use, ideally use pipx:

pipx install utk_exodus

This will install the application in a virtual environment and make it available to you where ever you are in your path, so that you can use it from anywhere without needing to understand the intricacies of Python.

If you don't want to use pipx, you can install the whole library with the following command but do so mindfully:

pip install utk_exodus

Before You Start

Before you start, you need to have a few things in place:

  1. Exodus assumes you have the following environmental variables set appropriately:
    • FEDORA_USER: this is a user with read access to the Fedora repository
    • FEDORA_PASSWORD: the password for that Fedora user
    • FEDORA_URI: the base URI for where Fedora is installed
  2. If you're looking for these values, you can find them in the Exodus Environment of this repository in Settings.

Using

There are several interfaces for the application.

You can always find out what interfaces exist with:

exodus --help

Similarly, you can get help for a specific interface with:

exodus <interface> --help

If you want to get works and files, and you have metadata files, use:

exodus works_and_files --path /path/to/metadata -o /path/to/directory/to/store/files

If you want to get works and files, and you don't have metadata files, you need to specify a collection and a work type:

exodus works_and_files --collection "namespace:identifier" --model book -o /path/to/output/directory

If you just want works, use:

exodus works --path /path/to/metadata

If for some reason you need to create a files sheet for works after the fact, use:

exodus add_files --sheet path/to/sheet.csv --files_sheet path/to/files_sheet.csv 

If you need to remove unused values from an import or work sheet, you can:

exodus remove_old_values --sheet path/to/sheet.csv --old_sheet path/to/old_sheet.csv --new_sheet path/to/new_sheet.csv

If you want to generate a full template for a metadata import, use:

exodus generate_template --model book -o /path/to/sheet.csv

If you want to generate a sheet of checksums for files that failed to import, you can:

exodus hash_errors --path /path/to/directory --output /path/to/sheet.csv

If you want to generate an import sheet for all collections, you can:

exodus generate_collection_metadata

If you want to generate an import sheet for a single collection, you can:

exodus generate_collection_metadata --collection "namespace:identifier"

What's Missing Here Right Now

  • The ability to create pcdm:Collection objects.
  • The ability to create a new metadata import from a previous import

Understanding Configs

Exodus migrates works and filesets according to the UTK Metadata mapping. To do this, Exodus uses yml files for migration. By default, exodus treats everything agnostically and relies on the xpaths section of the base mapping to determine how a concept is mapped. If a property (or properties) have complex rules, a class can be written to handle the special case. When this happens, the yml should have a special property, and it should be defined in MetadataMapping().__lookup_special_property().

An agnostic property should look like this in the yml:

  - name: table_of_contents
    xpaths:
      - 'mods:tableOfContents'
    property: "http://purl.org/dc/terms/tableOfContents"

A complex property might look like this:

  - name: title_and_alternative_title
    xpaths:
      - 'mods:titleInfo[not(@supplied)]/mods:title'
      - 'mods:titleInfo[@supplied="yes"]/mods:title'
    properties:
      - "http://purl.org/dc/terms/title"
      - "http://purl.org/dc/terms/alternative"
    special: "TitleProperty"

An agnostic property must always have the property property while a complex property may have property or properties.

Development

Running Tests

Tests can be run with Pytest:

pytest

New versions of packages can be published and pushed to pypi.org with Poetry.

First, make sure you follow semantic versioning and set a new release version in pyproject.toml:

[tool.poetry]
name = "utk-exodus"
version = "0.2.1"

Warning: Failure to follow semantic versioning may break installations and updates with pip and pipx.

After you set a new version in pyproject.toml, build an publish your new package like so:

poetry build
poetry publish

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

utk_exodus-0.2.5.tar.gz (35.0 kB view details)

Uploaded Source

Built Distribution

utk_exodus-0.2.5-py3-none-any.whl (46.5 kB view details)

Uploaded Python 3

File details

Details for the file utk_exodus-0.2.5.tar.gz.

File metadata

  • Download URL: utk_exodus-0.2.5.tar.gz
  • Upload date:
  • Size: 35.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.12.7 Darwin/23.6.0

File hashes

Hashes for utk_exodus-0.2.5.tar.gz
Algorithm Hash digest
SHA256 232cf17b8992a42aff40c3472603aba57550df0f587f7c1394d118f34d5d7632
MD5 ca81fb8fc65aec81a5792f8adc2911b4
BLAKE2b-256 b5c0ff57d271e8cbda1fd0d4323cbddc8bd787244f6bae2002a9bbcba53a76f1

See more details on using hashes here.

File details

Details for the file utk_exodus-0.2.5-py3-none-any.whl.

File metadata

  • Download URL: utk_exodus-0.2.5-py3-none-any.whl
  • Upload date:
  • Size: 46.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.8.3 CPython/3.12.7 Darwin/23.6.0

File hashes

Hashes for utk_exodus-0.2.5-py3-none-any.whl
Algorithm Hash digest
SHA256 ddcfc229d132596517a23b8671ed79d6f2dcf91a75c849652eba4c9f3d76bc2e
MD5 94be23c2bf3dc7a6e5082b129b0cc3a1
BLAKE2b-256 25c2242fa471098bd0fd17713c320bc47ab55a51d04353fbe5145061016c1390

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page