Skip to main content

🐳 Ocean/Web3py wrapper.

Project description



💧 Python wrapper, allowing to integrate the basic Ocean/ capabilities

PyPI Travis (.com) GitHub contributors Codacy Badge Codacy coverage

Table of Contents


Squid-py include the methods to make easy the connection with contracts deployed in different networks. This repository include also the methods to encrypt and decrypt information using the Parity Secret Store.


Python 3.6


Install Squid:

pip install squid-py


import os
import time

from squid_py import (

# Make a new instance of Ocean
ocean = Ocean()
config = ConfigProvider.get_config()
# make account instance, assuming the ethereum account and password are set 
# in the config file `config.ini`
account = Account(config.parity_address, config.parity_password)

# Let's start by registering an asset in the Ocean network
metadata = Metadata.get_example()

# purchase and service endpoints require `brizo.url` is set in the config file
# or passed to Ocean instance in the config_dict.
# define the services to include in the new asset DDO

ddo = ocean.register_asset(metadata, account)

# Now we have an asset registered, we can verify it exists by resolving the did
_ddo = ocean.resolve_asset_did(ddo.did)
# ddo and _ddo should be identical

# search for assets
asset_ddo = ocean.search_assets_by_text('Ocean protocol')[0]
# Need some ocean tokens to be able to purchase assets

# Start the purchase/consume request. This will automatically make a payment from the specified account.
account = Account(config.parity_address, config.parity_password)
service_agreement_id = ocean.purchase_asset(asset_ddo.did, 0, account)

# after a short wait (seconds to minutes) the asset data files should be available in the `downloads.path` defined in config
# wait a bit to let things happen

# Asset files are saved in a folder named after the asset id
dataset_dir = os.path.join(ConfigProvider.get_config().downloads_path, f'datafile.{asset_ddo.asset_id}.0')
if os.path.exists(dataset_dir):
    print('asset files downloaded: {}'.format(os.listdir(dataset_dir)))


config_dict = {
    'keeper-contracts': {
        # Point to an Ethereum RPC client. Note that Squid learns the name of the network to work with from this client.
        'keeper.url': 'http://localhost:8545',
        # Specify the keeper contracts artifacts folder (has the smart contracts definitions json files). When you
        # install the package, the artifacts are automatically picked up from the `keeper-contracts` Python
        # dependency unless you are using a local ethereum network.
        'keeper.path': 'artifacts',
        'secret_store.url': 'http://localhost:12001',
        'parity.url': 'http://localhost:8545',
        'parity.address': '',
        'parity.password': '',

    'resources': {
        # aquarius is the metadata store. It stores the assets DDO/DID-document
        'aquarius.url': 'http://localhost:5000',
        # Brizo is the publisher's agent. It serves purchase and requests for both data access and compute services
        'brizo.url': 'http://localhost:8030',
        # points to the local database file used for storing temporary information (for instance, pending service agreements).
        'storage.path': 'squid_py.db',
        # Where to store downloaded asset files
        'downloads.path': 'consume-downloads'

In addition to the configuration file, you may use the following environment variables (override the corresponding configuration file values):



  1. Set up a virtual environment

    virtualenv venv -p python3.6
    source venv/bin/activate 
  2. Install requirements

    pip install -r requirements_dev.txt
  3. Create the local testing environment using barge. Once cloned that repository, you can start the cluster running:

    ./ --latest --no-brizo --no-pleuston --local-spree-node

    It runs an Aquarius node and an Ethereum RPC client. For details, read docker-compose.yml.

  4. Create local configuration file

    cp config.ini config_local.ini

    config_local.ini is used by unit tests.

  5. Copy keeper artifacts

    A bash script is available to copy keeper artifacts into this file directly from a running docker image. This script needs to run in the root of the project. The script waits until the keeper contracts are deployed, and then copies the artifacts.


    The artifacts contain the addresses of all the deployed contracts and their ABI definitions required to interact with them.

  6. Run the automated tests

    TEST_TIER=<tier> python3 test # where <tier> is unit, integration or e2e, depending on how much of the tests you want to run (e2e runs everything)
  7. Run tests automatically as you change code while doing TDD

    TEST_TIER=<tier> ptw

Code style

The information about code style in python is documented in this two links python-developer-guide and python-style-guide.


Automatic tests are setup via Travis, executing tox. Our test use pytest framework.

New Version / New Release



Copyright 2018 Ocean Protocol Foundation Ltd.

Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at

Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
See the License for the specific language governing permissions and
limitations under the License.

Project details

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Filename, size & hash SHA256 hash help File type Python version Upload date
squid_py-0.2.21-py2.py3-none-any.whl (85.9 kB) Copy SHA256 hash SHA256 Wheel py2.py3
squid-py-0.2.21.tar.gz (116.2 kB) Copy SHA256 hash SHA256 Source None

Supported by

Elastic Elastic Search Pingdom Pingdom Monitoring Google Google BigQuery Sentry Sentry Error logging AWS AWS Cloud computing DataDog DataDog Monitoring Fastly Fastly CDN SignalFx SignalFx Supporter DigiCert DigiCert EV certificate StatusPage StatusPage Status page