Skip to main content

GoodData Cloud lifecycle automation pipelines

Project description

GoodData Pipelines

A high-level library for automating the lifecycle of GoodData Cloud (GDC).

You can use the package to manage following resources in GDC:

  1. Provisioning (create, update, delete)
    • User profiles
    • User Groups
    • User/Group permissions
    • User Data Filters
    • Child workspaces (incl. Workspace Data Filter settings)
  2. Backup and restore of workspaces
    • Create and backup snapshots of workspace metadata to local storage, AWS S3, or Azure Blob Storage
  3. LDM Extension
    • extend the Logical Data Model of a child workspace with custom datasets and fields

In case you are not interested in incorporating a library in your own program but would like to use a ready-made script, consider having a look at GoodData Productivity Tools.

Provisioning

The entities can be managed either in full load or incremental way.

Full load means that the input data should represent the full and complete desired state of GDC after the script has finished. For example, you would include specification of all child workspaces you want to exist in GDC in the input data for workspace provisioning. Any workspaces present in GDC and not defined in the source data (i.e., your input) will be deleted.

On the other hand, the incremental load treats the source data as instructions for a specific change, e.g., a creation or a deletion of a specific workspace. You can specify which workspaces you would want to delete or create, while the rest of the workspaces already present in GDC will remain as they are, ignored by the provisioning script.

The provisioning module exposes Provisioner classes reflecting the different entities. The typical usage would involve importing the Provisioner class and the data input data model for the class and planned provisioning method:

import os
import logging

from csv import DictReader
from pathlib import Path

# Import the Entity Provisioner class and corresponding model from the gooddata_pipelines library
from gooddata_pipelines import UserFullLoad, UserProvisioner

# Create the Provisioner instance - you can also create the instance from a GDC yaml profile
provisioner = UserProvisioner(
    host=os.environ["GDC_HOSTNAME"], token=os.environ["GDC_AUTH_TOKEN"]
)

# Optional: set up logging and subscribe to logs emitted by the provisioner
logging.basicConfig(level=logging.INFO)
logger = logging.getLogger(__name__)
provisioner.logger.subscribe(logger)

# Load your data from your data source
source_data_path: Path = Path("path/to/some.csv")
source_data_reader = DictReader(source_data_path.read_text().splitlines())
source_data = [row for row in source_data_reader]

# Validate your input data
full_load_data: list[UserFullLoad] = UserFullLoad.from_list_of_dicts(
    source_data
)

# Run the provisioning
provisioner.full_load(full_load_data)

Ready-made scripts covering the basic use cases can be found here in the GoodData Productivity Tools repository.

Backup and Restore of Workspaces

The backup and restore module allows you to create snapshots of GoodData Cloud workspaces and restore them later. Backups can be stored locally, in AWS S3, or Azure Blob Storage.

import os

from gooddata_pipelines import BackupManager
from gooddata_pipelines.backup_and_restore.models.storage import (
    BackupRestoreConfig,
    LocalStorageConfig,
    StorageType,
)

# Configure backup storage
config = BackupRestoreConfig(
    storage_type=StorageType.LOCAL,
    storage=LocalStorageConfig(),
)

# Create the BackupManager instance
backup_manager = BackupManager.create(
    config=config,
    host=os.environ["GDC_HOSTNAME"],
    token=os.environ["GDC_AUTH_TOKEN"]
)

# Backup specific workspaces
backup_manager.backup_workspaces(workspace_ids=["workspace1", "workspace2"])

# Backup workspace hierarchies (workspace + all children)
backup_manager.backup_hierarchies(workspace_ids=["parent_workspace"])

# Backup entire organization
backup_manager.backup_entire_organization()

For S3 or Azure Blob Storage, configure the appropriate storage type and credentials in BackupRestoreConfig.

Bugs & Requests

Please use the GitHub issue tracker to submit bugs or request features.

Changelog

See GitHub releases for released versions and a list of changes.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gooddata_pipelines-1.63.1.dev2.tar.gz (96.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

gooddata_pipelines-1.63.1.dev2-py3-none-any.whl (98.6 kB view details)

Uploaded Python 3

File details

Details for the file gooddata_pipelines-1.63.1.dev2.tar.gz.

File metadata

  • Download URL: gooddata_pipelines-1.63.1.dev2.tar.gz
  • Upload date:
  • Size: 96.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.12

File hashes

Hashes for gooddata_pipelines-1.63.1.dev2.tar.gz
Algorithm Hash digest
SHA256 16a6fa39cb5fcb7b6fa666cc970ebbf44e992e98d6ef22e091ff58dfc4160070
MD5 df0c5ec7159b4fa71b38c6047cee72fe
BLAKE2b-256 993d98bec4b9964066490e0d40cf3338740d92f2a7f526422a1c308a04976ca6

See more details on using hashes here.

Provenance

The following attestation bundles were made for gooddata_pipelines-1.63.1.dev2.tar.gz:

Publisher: dev-release.yaml on gooddata/gooddata-python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file gooddata_pipelines-1.63.1.dev2-py3-none-any.whl.

File metadata

File hashes

Hashes for gooddata_pipelines-1.63.1.dev2-py3-none-any.whl
Algorithm Hash digest
SHA256 eac27ad0db2fe2f342b5e1c58d70d51c439b71720f5f1858787a5b062c8eba50
MD5 ac988792ed4462b1473afa28756bbf3d
BLAKE2b-256 57c7728a39ea078958685c5cb6f3ddb70fe2726e2b0d438ad1b5687ba6ed1059

See more details on using hashes here.

Provenance

The following attestation bundles were made for gooddata_pipelines-1.63.1.dev2-py3-none-any.whl:

Publisher: dev-release.yaml on gooddata/gooddata-python-sdk

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page