Skip to main content

Aggregation code for Zooniverse panoptes projects.

Project description

Getting started

This is a collection of external reducers written for caesar and offline use.


Documentation

You can find the latest documentation on the aggregations code's website.


Installing for offline use

With your own python install (python 3 only)

Instal the latest stable release:

pip install panoptes_aggregation

Or for development or testing, you can install the development version directly from GitHub:

pip install -U git+git://github.com/zooniverse/aggregation-for-caesar.git

Upgrade and existing installation:

pip install -U panoptes_aggregation

With Docker

https://docs.docker.com/get-started/

Using docker-compose https://docs.docker.com/compose/

docker-compose -f docker-compose.local_scripts.yml build local_scripts

From the root directory of this repository, run the desired python scripts using the docker image, e.g. config_workflow_panoptes --help

docker-compose -f docker-compose.local_scripts.yml run --rm local_scripts config_workflow_panoptes --help

Or directly via docker

docker build . -f Dockerfile.bin_cmds -t aggregation_for_caesar

From the root directory of this repository, run the desired python scripts using the docker image, e.g. config_workflow_panoptes --help

docker run -it --rm --name config_workflow_panoptes -v "$PWD":/usr/src/aggregation aggregation_for_caesar config_workflow_panoptes --help

Installing for online use

The docker file included is ready to be deployed on any server. Once deployed, the extractors will be available on the /extractors/<name of extractor function> routes and the reducers will be available on the /reducers/<name of reducer function> routes. Any keywords passed into these functions should be included as url parameters on the route (e.g. https://aggregation-caesar.zooniverse.org/extractors/point_extractor_by_frame?task=T0). For more complex keywords (e.g. detals for subtasks), python's urllib.parse.urlencode can be used to translate a keyword list into the proper url encoding.

The documentation will be built and available on the /docs route.

Build/run the app in docker locally

To run a local version use:

docker-compose build
docker-compose up

and listen on localhost:5000.

Running tests in the docker container

To run the tests use:

docker-compose run --rm aggregation nosetests

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

panoptes_aggregation-1.2.0.tar.gz (62.2 kB view details)

Uploaded Source

Built Distribution

panoptes_aggregation-1.2.0-py3-none-any.whl (113.1 kB view details)

Uploaded Python 3

File details

Details for the file panoptes_aggregation-1.2.0.tar.gz.

File metadata

  • Download URL: panoptes_aggregation-1.2.0.tar.gz
  • Upload date:
  • Size: 62.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.18.4 setuptools/40.1.1 requests-toolbelt/0.8.0 tqdm/4.25.0 CPython/3.5.2

File hashes

Hashes for panoptes_aggregation-1.2.0.tar.gz
Algorithm Hash digest
SHA256 07f585672d0672794843bb4fec36692c8834470bcf3614e678aadd60829064cb
MD5 bb23c343f18801055fdf572eac6ef8e4
BLAKE2b-256 e3dc5308eb9a2c3e61e88220704fa3d004106d6b577d0f6012370221a7e45dd6

See more details on using hashes here.

File details

Details for the file panoptes_aggregation-1.2.0-py3-none-any.whl.

File metadata

  • Download URL: panoptes_aggregation-1.2.0-py3-none-any.whl
  • Upload date:
  • Size: 113.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.18.4 setuptools/40.1.1 requests-toolbelt/0.8.0 tqdm/4.25.0 CPython/3.5.2

File hashes

Hashes for panoptes_aggregation-1.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 4112e3c86e020d327460d2b39195074bb5493c3f50f159618d678d32e745ffc2
MD5 7f2808d3cb82d32eae3fb6b25fc93d66
BLAKE2b-256 aaa80bb364f7929fa3df9ea016176d03dfcbdeeccd59d77052a74ef4e84463c4

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page