Skip to main content

Aggregation code for Zooniverse panoptes projects.

Project description

Getting started

This is a collection of external reducers written for caesar and offline use.


Documentation

You can find the latest documentation on the aggregations code's website.


Installing for offline use

With your own python install (python 3 only)

Instal the latest stable release:

pip install panoptes_aggregation

Or for development or testing, you can install the development version directly from GitHub:

pip install -U git+git://github.com/zooniverse/aggregation-for-caesar.git

Upgrade and existing installation:

pip install -U panoptes_aggregation

With Docker

https://docs.docker.com/get-started/

Using docker-compose https://docs.docker.com/compose/

docker-compose -f docker-compose.local_scripts.yml build local_scripts

From the root directory of this repository, run the desired python scripts using the docker image, e.g. config_workflow_panoptes --help

docker-compose -f docker-compose.local_scripts.yml run --rm local_scripts config_workflow_panoptes --help

Or directly via docker

docker build . -f Dockerfile.bin_cmds -t aggregation_for_caesar

From the root directory of this repository, run the desired python scripts using the docker image, e.g. config_workflow_panoptes --help

docker run -it --rm --name config_workflow_panoptes -v "$PWD":/usr/src/aggregation aggregation_for_caesar config_workflow_panoptes --help

Installing for online use

The docker file included is ready to be deployed on any server. Once deployed, the extractors will be available on the /extractors/<name of extractor function> routes and the reducers will be available on the /reducers/<name of reducer function> routes. Any keywords passed into these functions should be included as url parameters on the route (e.g. https://aggregation-caesar.zooniverse.org/extractors/point_extractor_by_frame?task=T0). For more complex keywords (e.g. detals for subtasks), python's urllib.parse.urlencode can be used to translate a keyword list into the proper url encoding.

The documentation will be built and available on the /docs route.

Build/run the app in docker locally

To run a local version use:

docker-compose build
docker-compose up

and listen on localhost:5000.

Running tests in the docker container

To run the tests use:

docker-compose run --rm aggregation nosetests

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

panoptes_aggregation-1.3.0.tar.gz (64.8 kB view details)

Uploaded Source

Built Distribution

panoptes_aggregation-1.3.0-py3-none-any.whl (119.0 kB view details)

Uploaded Python 3

File details

Details for the file panoptes_aggregation-1.3.0.tar.gz.

File metadata

  • Download URL: panoptes_aggregation-1.3.0.tar.gz
  • Upload date:
  • Size: 64.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.18.4 setuptools/40.1.1 requests-toolbelt/0.8.0 tqdm/4.25.0 CPython/3.5.2

File hashes

Hashes for panoptes_aggregation-1.3.0.tar.gz
Algorithm Hash digest
SHA256 9939990a3d4a93429bbf0b4de83ef45aa4b0d50e55414b5b89a3b5fefca802ef
MD5 063ebf4fa882d401083b76940b736d3a
BLAKE2b-256 a3900c7208831dfd17b999c134a33217c0cc81118733b91d838cc59d2716e4d3

See more details on using hashes here.

File details

Details for the file panoptes_aggregation-1.3.0-py3-none-any.whl.

File metadata

  • Download URL: panoptes_aggregation-1.3.0-py3-none-any.whl
  • Upload date:
  • Size: 119.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.18.4 setuptools/40.1.1 requests-toolbelt/0.8.0 tqdm/4.25.0 CPython/3.5.2

File hashes

Hashes for panoptes_aggregation-1.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 ad3acc7daf1c1d2da24c8ea6c74f79593a089f768d1f673a35a9906c5d7b12f6
MD5 fbd77f27fc09c06e75b8f73249ecac5f
BLAKE2b-256 4adca16e672b945ba548a8feb91a2e07d3ae209ae78f20e3f095fce6b3be792a

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page