Skip to main content

A Python SDK for interacting with Coherent Spark APIs

Project description

Coherent Spark Python SDK

PyPI version

The Coherent Spark Python SDK (currently in Beta) is designed to elevate the developer experience and provide convenient access to Coherent Spark APIs.

👋 Just a heads-up:

This SDK is supported by the community. If you encounter any bumps while using it, please report them here by creating a new issue.

Installation

pip install -U cspark

🫣 This Python library requires Python 3.7+.

Usage

To use the SDK, you need a Coherent Spark account that lets you access the following:

  • User authentication (API key, bearer token, or OAuth2.0 client credentials details)
  • Base URL (including the environment and tenant name)
  • Spark service URI (to locate a specific resource):
    • folder - the folder name (where the service is located)
    • service - the service name
    • version - the semantic version a.k.a revision number (e.g., 0.4.2)

A folder contains one or more services, which can have multiple versions. Technically speaking, when you're operating with a service, you're actually interacting with a specific version of that service (the latest version by default - unless specified otherwise).

Hence, there are various ways to indicate a Spark service URI:

  • {folder}/{service}[?{version}] - version is optional.
  • service/{service_id}
  • version/{version_id}

IMPORTANT: Avoid using URL-encoded characters in the service URI.

Here's an example of how to execute a Spark service:

import cspark.sdk as Spark

spark = Spark.Client(env='my-env', tenant='my-tenant', api_key='my-api-key')
with spark.services as services:
    response = services.execute('my-folder/my-service', inputs={'value': 42})
    print(response.data)

Explore the examples and documentation folders to find out more about the SDK's capabilities.

PRO TIP: A service URI locator can be combined with other parameters to locate a specific service (or version of it) when it's not a string. For example, you may execute a public service using a UriParams object by specifying the folder, service, and public properties.

import cspark.sdk as Spark

spark = Spark.Client(env='my-env', tenant='my-tenant', api_key='open')

with spark.services as services:
    uri = Spark.UriParams(folder='my-folder', service='my-service', public=True)
    response = services.execute(uri, inputs={'value': 42})
    print(response.data)

# The final URI in this case is:
#    'my-tenant/api/v3/public/folders/my-folder/services/my-service/execute'

See the Uri and UriParams classes for more details.

Client Options

As shown in the examples above, the Spark.Client is your entry point to the SDK. It is quite flexible and can be configured with the following options:

Base URL

base_url (default: os.getenv['CSPARK_BASE_URL']): indicates the base URL of Coherent Spark APIs. It should include the tenant and environment information.

spark = Spark.Client(base_url='https://excel.my-env.coherent.global/my-tenant')

Alternatively, a combination of env and tenant options can be used to construct the base URL.

spark = Spark.Client(env='my-env', tenant='my-tenant')

Authentication

The SDK supports three types of authentication schemes:

  • api_key (default: os.getenv['CSPARK_API_KEY']): indicates the API key (also known as synthetic key), which is sensitive and should be kept secure.
spark = Spark.Client(api_key='my-api-key')

PRO TIP: The Spark platform supports public APIs that can be accessed without any form of authentication. In that case, you need to set api_key to open in order to create a Spark.Client.

  • token (default: os.getenv['CSPARK_BEARER_TOKEN']): indicates the bearer token. It can be prefixed with 'Bearer' or not. A bearer token is usually valid for a limited time and should be refreshed periodically.
spark = Spark.Client(token='Bearer my-access-token') # with prefix
# or
spark = Spark.Client(token='my-access-token') # without prefix
  • oauth (default: os.getenv['CSPARK_CLIENT_ID'] and os.getenv['CSPARK_CLIENT_SECRET'] or os.getenv['CSPARK_OAUTH_PATH']): indicates the OAuth2.0 client credentials. You can either provide the client ID and secret directly or the file path to the JSON file containing the credentials.
spark = Spark.Client(oauth={'client_id': 'my-client-id', 'client_secret': 'my-client-secret'})
# or
spark = Spark.Client(oauth='path/to/oauth/credentials.json')

Additional Settings

  • timeout (default: 60000 ms): indicates the maximum amount of time that the client should wait for a response from Spark servers before timing out a request.

  • max_retries (default: 2): indicates the maximum number of times that the client will retry a request in case of a temporary failure, such as an unauthorized response or a status code greater than 400.

  • retry_interval (default: 1 second): indicates the delay between each retry.

  • logger (default: True): enables or disables logs for the SDK.

spark = Spark.Client(logger=False)

Client Errors

SparkError is the base class for all custom errors thrown by the SDK. There are two types of it:

  • SparkSdkError: usually thrown when an argument (user entry) fails to comply with the expected format. Because it's a client-side error, it will include the invalid entry as the cause in most cases.
  • SparkApiError: when attempting to communicate with the API, the SDK will wrap any sort of failure (any error during the roundtrip) into SparkApiError, which includes the HTTP status code of the response and the request_id, a unique identifier of the request.

Some of the derived SparkApiError are:

Type Status When
InternetError 0 no internet access
BadRequestError 400 invalid request
UnauthorizedError 401 missing or invalid credentials
ForbiddenError 403 insufficient permissions
NotFoundError 404 resource not found
ConflictError 409 resource already exists
RateLimitError 429 too many requests
InternalServerError 500 server-side error
ServiceUnavailableError 503 server is down
UnknownApiError None unknown error

API Parity

The SDK aims to provide full parity with the Spark APIs over time. Below is a list of the currently supported APIs.

Authentication API - manages access tokens using OAuth2.0 Client Credentials flow:

  • Authorization.oauth.retrieve_token(config) generates new access tokens.

Services API - manages Spark services:

  • Spark.services.execute(uri, inputs) executes a Spark service.
  • Spark.services.get_versions(uri) lists all the versions of a service.
  • Spark.services.get_schema(uri) gets the schema of a service.
  • Spark.services.get_metadata(uri) gets the metadata of a service.

Batches API - manages asynchronous batch processing:

  • Spark.batches.describe() describes the batch pipelines across a tenant.
  • Spark.batches.create(params, [options]) creates a new batch pipeline.
  • Spark.batches.of(id) defines a client-side batch pipeline by ID.
  • Spark.batches.of(id).get_info() gets the details of a batch pipeline.
  • Spark.batches.of(id).get_status() gets the status of a batch pipeline.
  • Spark.batches.of(id).push(data, [options]) adds input data to a batch pipeline.
  • Spark.batches.of(id).pull([options]) retrieves the output data from a batch pipeline.
  • Spark.batches.of(id).dispose() closes a batch pipeline.
  • Spark.batches.of(id).cancel() cancels a batch pipeline.

Contributing

Feeling motivated enough to contribute? Great! Your help is always appreciated.

Please read CONTRIBUTING.md for details on the code of conduct, and the process for submitting pull requests.

Copyright and License

Apache-2.0

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cspark-0.1.3.tar.gz (22.0 kB view details)

Uploaded Source

Built Distribution

cspark-0.1.3-py3-none-any.whl (27.2 kB view details)

Uploaded Python 3

File details

Details for the file cspark-0.1.3.tar.gz.

File metadata

  • Download URL: cspark-0.1.3.tar.gz
  • Upload date:
  • Size: 22.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.12.3

File hashes

Hashes for cspark-0.1.3.tar.gz
Algorithm Hash digest
SHA256 2ce6c62362c94bfb52fcd5cfb9cadd1ffa1fef98db75bf64e1967e13effe2d1a
MD5 6591d95f0299fe139f8cd6bedb875738
BLAKE2b-256 8b0d649fa4782cc097dd243cfea67b981c25a8473daf6d0128e0358c949db650

See more details on using hashes here.

File details

Details for the file cspark-0.1.3-py3-none-any.whl.

File metadata

  • Download URL: cspark-0.1.3-py3-none-any.whl
  • Upload date:
  • Size: 27.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.12.3

File hashes

Hashes for cspark-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 4fdd6c92186224931f909cf354c4fc56d70de12b1fbbfd7e8f7a2e329441a82e
MD5 b124a8b826f9542e97891600d1a3d263
BLAKE2b-256 37d2c103ac4d5654ae95986a69c8d714bc2916225140ea1ff5a20f42d17ce4a6

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page