Skip to main content

No project description provided

Project description

linkml-transformer

LinkML Transformer is a framework for specifying and executing mappings between data models.

Features:

  • YAML-based lightweight syntax
  • Python library for executing mappings on data files
  • Ability to compile to other frameworks (forthcoming)
  • Derivation of target (implicit) schemas, allowing easy customization of data models (profiling)
  • Simple YAML dictionaries for simple mappings
  • Automatic unit conversion
  • Use of subset of Python to specify complex mappings
  • Visualizations of mappings
  • Mappings are reversible (provided all expressions used are reversible)

Status:

The transformation data model is not yet fully stable, and may be subject to change. Not all parts of the model are implemented in the reference Python framework.

Quickstart

Basic idea

Given an object that conforms to a LinkML schema,e.g.:

given_name: Jane
family_name: Doe
height_in_cm: 172.0
age_in_years: 33
aliases: [Janey, Janie]

Define a mapping in YAML:

class_derivations:
  Individual:
    populated_from: Person
    slot_derivations:
      name:
        expr: "{given_name} + ' ' + {family_name}"
        description: Concatenating given and family names
          note this is a bad assumption for names in general,
          this is just for demonstration
      height_in_meters:
        populated_from: height_in_cm
        unit_conversion:
          target_unit: m
      aliases:
        populated_from: aliases
        stringification:
          delimiter: '|'

Transform the data:

name: Jane Doe
height_in_meters: 1.72
aliases: Janey|Janie

Installation and usage

Installation and command line usage:

pip[x] install linkml-transformer
cd tests/input/examples/personinfo_basic
linkml-tr map-data \
  -T transform/personinfo-to-agent.transform.yaml \
  -s source/personinfo.yaml \
   data/Container-001.yaml \
   -o output/Translated-001.yaml

The command line has subcommands for:

  • map-data - map data from a source schema to a target schema
  • derive-schema - derive a target schema from a source schema and a mapping
  • invert - reverses a mapping
  • compile - compiles a mapping to another framework
    • markdown - for generating static sites
    • graphviz - for generating visualizations
    • python - (partial)
    • forthcoming: r2rml, ...

Details

This repo contains both:

  • A data model for a data model transformation language
  • A reference python implementation

The transformation language is specified in terms of LinkML schemas. It is intended to be a ployglot transformation language, used for specifying how to map data models independent of underlying representation (TSVs, JSON/YAML, RDF, SQL Database, ...).

Use cases include:

  • ETL and mapping from one data model to another
  • Database migrations (one version of a schema to another)
  • Creating "profiles"
  • Specifying mappings between different serializations of a model (e.g. OO to Relational)
  • Mapping between normalized/non-redundant forms and denormalized/query-optimized forms

Tutorials

Data Model

The data model for transformations mirrors the data model for schemas:

  • A top level TransformationSpecification class contains:
    • Zero or more ClassDerivation objects, specifying how to map to a class, containing:
      • Zero or more SlotDerivation objects, specifying how to map to a slot, containing:
    • Zero or more EnumDerivation objects, specifying how to map permissable values.

See the generated docs

Conformance Suite

The Transformers conformance suite contains a collection of tests for each feature of the language.

See:

Running the code

linkml-tr --help
Usage: linkml-tr [OPTIONS] COMMAND [ARGS]...

  CLI for linkml-transformer.

Options:
  -v, --verbose
  -q, --quiet TEXT
  --help            Show this message and exit.

Commands:
  derive-schema  Derive a schema from a source schema and a mapping.
  map-data       Map data in a source schema using a transformation.

map-data

Transforms (maps) data from a source schema to a target schema. This could range from a simple data dictionary mapping through to a complex mappings.

cd tests/input/examples/personinfo_basic
linkml-tr map-data -T transform/personinfo-to-agent.transform.yaml -s source/personinfo.yaml  data/Container-001.yaml

derive-schema

cd tests/input/examples/personinfo_basic
linkml-tr derive-schema -T transform/personinfo-to-agent.transform.yaml source/personinfo.yaml

Examples

See the tests folder for most up-to-date examples

Mapping between two similar data models

Given a source object

persons:
  - id: P:001
    name: fred bloggs
    primary_email: fred.bloggs@example.com
    age_in_years: 33
    has_familial_relationships:
      - type: SIBLING_OF
        related_to: P:002
    current_address:
      street: 1 oak street
    aliases:
      - a
      - b
    has_medical_history:
      - diagnosis:
          id: C:001
          name: c1
      - diagnosis:
          id: C:002
  - id: P:002
    name: Alison Wu
    has_familial_relationships:
      - type: SIBLING_OF
        related_to: P:001
    has_medical_history:
      - diagnosis:
          id: C:001
          name: c1 (renamed)
organizations:
  - id: ROR:1
    name: Acme

and a corresponding schema, consider the case of mapping to a largely isomorphic schema, with some minor differences:

  • class names are changes (e.g Person to Agent)
  • age is represented as a string, e.g. "33 years"
  • some fields are denormalized

The mapping may look like:

id: my-mappings
title: my mappings
prefixes:
  foo: foo
source_schema: s1
target_schema: s2
class_derivations:
  Container:
    populated_from: Container
    slot_derivations:
      agents:
        populated_from: persons
  Agent:
    populated_from: Person
    slot_derivations:
      
      ## implicitly same name in Agent
      id:

      label:
        populated_from: name

      age:
        expr: "str({age_in_years})+' years'"

      primary_email:

      gender:

      has_familial_relationships:
        populated_from: has_familial_relationships

  FamilialRelationship:
    populated_from: FamilialRelationship
    slot_derivations:
      type:
      related_to:

Deriving Schemas

Formally a mapping consists of a source schema S, a target schema T, and a mapping M.

In practice, any of these can be partially specified and derived from the others.

For example:

  • given S, and T, derive isomorphic mappings M based on shared URIs
  • given S and M, derive T` by applying M to the schema

You can use the derive-schema command to derive a schema from a source schema and a mapping. This can also be thought of as "profiling" a schema (in the FHIR sense).

See tests/test_mapper/test_schema_mapper.py for examples

Measurements and units

- id: P:001
  height:
    value: 172.0
    unit: cm

<==>

- id: P:001
  height_in_cm: 172.0

<==>

- id: P:001
  height: "172.0 cm"

Not supported

  • Aggregation

Why another framework?

There are a number of excellent frameworks for performing mapping and data transformations. The LinkML Transformer framework was born out of a need for a framework that:

  • was not inherently tied to:
    • a particular serialization format (e.g. RDF, JSON, ...)
    • a particular programming language (e.g. Python, Java, ...)
    • a particular database system or database language (e.g. PostgreSQL or SQL or SPARQL)
    • not tied to a particular kind of transformation (e.g. ORM or Tables to RDF)
  • was a natural fit for the LinkML data modeling framework
  • was declarative and easy to perform machine reasoning over
  • is simple for simple use cases

In its current state, this framework is less powerful and expressive than many other frameworks or methodologies for performing transformations. If you need to perform complex data transformations, you might be better off using an expressive query language like SPARQL or SQL, or even just coding transformations directly in a programming language or library like Python or Pandas (but note that even for the coding use case, the LinkML Transformer framework can be useful as a standard way of documenting transformations).

Currently the main use case for this framework is mostly isomorphic transformations, with lightweight manipulations. These lend themselves well to a declarative framework. Uses cases that are a particularly good fit involve mapping between data-dictionary like standards, with large numbers of metadata elements, where these elements can often be mapped one-to-one, or with simple manipulations (e.g. unit conversions).

The origins lie partly in the SSSOM standard, which is intended as an ultra-simple way of specifying precise mappings between entities (e.g. schema.org Person is the same as DBPedia person). We observed that many people wanted to extend the language to perform more complex mappings. To address this, we help a workshop at the Biocuration conference in Padua, Italy in 2022.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

linkml_transformer-0.2.0.tar.gz (32.8 kB view details)

Uploaded Source

Built Distribution

linkml_transformer-0.2.0-py3-none-any.whl (42.9 kB view details)

Uploaded Python 3

File details

Details for the file linkml_transformer-0.2.0.tar.gz.

File metadata

  • Download URL: linkml_transformer-0.2.0.tar.gz
  • Upload date:
  • Size: 32.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.8.18

File hashes

Hashes for linkml_transformer-0.2.0.tar.gz
Algorithm Hash digest
SHA256 4402385249a8fd2209b8ed665d6bbfd315a6c6707f7ed19f1d5f58a5dd15a3ec
MD5 cc40833b741e22659429ce9087fa0831
BLAKE2b-256 32b9bc6b5f7d6ee505d036ac8f20108a0fea2c2eb3e494aecc747ef57a5fbc5e

See more details on using hashes here.

File details

Details for the file linkml_transformer-0.2.0-py3-none-any.whl.

File metadata

File hashes

Hashes for linkml_transformer-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 9f9382f7bdfe04e092e69b381bd63a60ebd79e588119e6b9c5944281bdc64eb1
MD5 ed60e4c72a3e9404e907809fed0c5227
BLAKE2b-256 e50ca910f3ef2f349bb036db8b0f8eb7311d1b1c2b990161a395b9f8e2709c6d

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page