Skip to main content

Aggregation code for Zooniverse panoptes projects.

Project description

Getting started

This is a collection of external reducers written for caesar and offline use.


Documentation

You can find the latest documentation on the aggregations code's website.


Installing for offline use

With your own python install (python 3 only)

Instal the latest stable release:

pip install panoptes_aggregation

Or for development or testing, you can install the development version directly from GitHub:

pip install -U git+git://github.com/zooniverse/aggregation-for-caesar.git

Upgrade and existing installation:

pip install -U panoptes_aggregation

With Docker

https://docs.docker.com/get-started/

Using docker-compose https://docs.docker.com/compose/

docker-compose -f docker-compose.local_scripts.yml build local_scripts

From the root directory of this repository, run the desired python scripts using the docker image, e.g. config_workflow_panoptes --help

docker-compose -f docker-compose.local_scripts.yml run --rm local_scripts config_workflow_panoptes --help

Or directly via docker

docker build . -f Dockerfile.bin_cmds -t aggregation_for_caesar

From the root directory of this repository, run the desired python scripts using the docker image, e.g. config_workflow_panoptes --help

docker run -it --rm --name config_workflow_panoptes -v "$PWD":/usr/src/aggregation aggregation_for_caesar config_workflow_panoptes --help

Installing for online use

The docker file included is ready to be deployed on any server. Once deployed, the extractors will be available on the /extractors/<name of extractor function> routes and the reducers will be available on the /reducers/<name of reducer function> routes. Any keywords passed into these functions should be included as url parameters on the route (e.g. https://aggregation-caesar.zooniverse.org/extractors/point_extractor_by_frame?task=T0). For more complex keywords (e.g. detals for subtasks), python's urllib.parse.urlencode can be used to translate a keyword list into the proper url encoding.

The documentation will be built and available on the /docs route.

Build/run the app in docker locally

To run a local version use:

docker-compose build
docker-compose up

and listen on localhost:5000.

Running tests in the docker container

To run the tests use:

docker-compose run --rm aggregation nosetests

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

panoptes_aggregation-1.2.1.tar.gz (64.2 kB view details)

Uploaded Source

Built Distribution

panoptes_aggregation-1.2.1-py3-none-any.whl (116.8 kB view details)

Uploaded Python 3

File details

Details for the file panoptes_aggregation-1.2.1.tar.gz.

File metadata

  • Download URL: panoptes_aggregation-1.2.1.tar.gz
  • Upload date:
  • Size: 64.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.18.4 setuptools/40.1.1 requests-toolbelt/0.8.0 tqdm/4.25.0 CPython/3.5.2

File hashes

Hashes for panoptes_aggregation-1.2.1.tar.gz
Algorithm Hash digest
SHA256 3636e000e78b53b7aa92f1da9342ee4777fb77c6cdad181be173a7e676aedd30
MD5 f06b1f4ddae73ffdaf60db89f5fd9793
BLAKE2b-256 5cf0ae2cbb55d037c886f7c925fa7da9b3c0c2e9b758059fd318fe0c6e780b55

See more details on using hashes here.

File details

Details for the file panoptes_aggregation-1.2.1-py3-none-any.whl.

File metadata

  • Download URL: panoptes_aggregation-1.2.1-py3-none-any.whl
  • Upload date:
  • Size: 116.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.18.4 setuptools/40.1.1 requests-toolbelt/0.8.0 tqdm/4.25.0 CPython/3.5.2

File hashes

Hashes for panoptes_aggregation-1.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 c52b895d2cb4cdb8bba6508a869286f3dd123fcc1035bbed73ba439e712dce40
MD5 2c9d877c867c49b5a64e91f126af1920
BLAKE2b-256 762122a62241ffea02f6ab33ab5c611df2c9084bb988cf428cf96734a854ed1c

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page