Skip to main content

No project description provided

Project description

=======

Bokchoi simplifies running Python batch jobs on AWS spot instances.
Bokchoi handles requesting spot instances, deploying your code and
ensures the spot requests are cancelled when all jobs are finished.

Getting Started
---------------

Installing
~~~~~~~~~~

To install bokchoi:

::

pip install bokchoi

Settings
~~~~~~~~

Say you have a project folder with a single python script:

::

YourProjectFolder
└─ deep_nn.py

In your project folder create a settings file named
**bokchoi_settings.json**:

::

YourProjectFolder
├─ deep_nn.py
└─ bokchoi_settings.json

This file should contain the following:

.. code:: json

{
"<project_name>": {
"EntryPoint": "deep_nn.main",
"Region": "us-east-1",
"Platform": "EC2",
"Requirements": [
"numpy==1.13.0",
"boto3==1.4.4"
],
"EC2": {
"InstanceCount": 1,
"SpotPrice": "0.1",
"LaunchSpecification": {
"ImageId": "ami-123456",
"InstanceType": "c4.large",
"SubnetId": "subnet-123456"
}
}
}
}

Deploying
~~~~~~~~~

Deploying your job to AWS is now as simple as running:

::

bokchoi project_name deploy

|
| Bokchoi will package your project and upload it to S3. You can then
use the following command to run your job:

::

bokchoi project_name run

|
| This will issue a spot request for the number of spot instances
specified in the settings file. Every spot instance will download the
packaged project from S3 and run the main function. Once the job is
complete the instance will shut down. When all instances are finished
the spot request will automatically be cancelled.

Undeploying
~~~~~~~~~~~

To undeploy your job, removing all resources from your AWS environment:

::

bokchoi project_name undeploy

|
| This will terminate any spot instances related to your job, cancel all
spot requests and remove the packaged project from S3. Any IAM
resources, such as policies, roles and instance profiles will also be
removed.

EMR
~~~

Bokchoi now also supports running python applications on Amazon EMR. To
run your app on an EMR cluster use the following settings:

.. code:: json

{
"<project_name>": {
"EntryPoint": "deep_nn.py",
"Region": "us-east-1",
"Platform": "EMR",
"Requirements": [
"numpy==1.13.0",
"boto3==1.4.4"
],
"EMR": {
"InstanceCount": 2,
"Version": "emr-5.8.0",
"SpotPrice": "0.10",
"LaunchSpecification": {
"InstanceType": "m1.medium",
"SubnetId": "subnet-123456",
"AdditionalSecurityGroups": ["sg-12ab34"]
}
}
}
}

Google Compute Engine
~~~~~~~~~~~~~~~~~~~~~

Google Compute Engine is also supported as a backend for python
applications. Simply change the EMR/EC2 part to the following:

The Google auth key can be obtained by creating a service account, which
can be created by following this guide:
https://cloud.google.com/iam/docs/creating-managing-service-account-keys

::

{
"bokchoi-gcp": {
"EntryPoint": "deep_nn.main",
"Platform": "GCP",
"Requirements": [
"numpy==1.13.0",
"boto3==1.4.4"
],
"GCP": {
"ProjectId": "google-project-id",
"AuthKeyLocation": "auth-key-user.json",
"Region": "europe-west4",
"Zone": "europe-west4-b",
"Bucket": "bokchoi-gcp",
"Network": "default",
"SubNetwork": "default",
"InstanceType": "n1-standard-1",
"Preemptible": true,
"DiskSizeGb": 25
}
}
}

+-----------------+-----------------+-----------------+-----------------+
| parameter | required | default | description |
+=================+=================+=================+=================+
| ProjectId | yes | None | Project id |
| | | | within Google |
| | | | Cloud |
+-----------------+-----------------+-----------------+-----------------+
| AuthKeyPath | yes | None | Path to the |
| | | | JSON or P12 |
| | | | auth file |
+-----------------+-----------------+-----------------+-----------------+
| Bucket | yes | None | Unique bucket |
| | | | name for Google |
| | | | Storage |
+-----------------+-----------------+-----------------+-----------------+
| Region | no | europe-west4 | Region where |
| | | | the instance |
| | | | will run |
+-----------------+-----------------+-----------------+-----------------+
| Zone | no | europe-west4-b | Zone where the |
| | | | instance will |
| | | | run |
+-----------------+-----------------+-----------------+-----------------+
| Network | no | default | Network where |
| | | | the instance |
| | | | will run |
+-----------------+-----------------+-----------------+-----------------+
| SubNetwork | no | default | Subnetwork |
| | | | where the |
| | | | instance will |
| | | | run |
+-----------------+-----------------+-----------------+-----------------+
| InstanceType | no | n1-standard-1 | Machine type |
+-----------------+-----------------+-----------------+-----------------+
| Preemptible | no | false | Whether the app |
| | | | runs on cheaper |
| | | | temporary |
| | | | instances |
+-----------------+-----------------+-----------------+-----------------+
| DiskSizeGb | no | 100 | Size (in GB) of |
| | | | the created |
| | | | disk |
+-----------------+-----------------+-----------------+-----------------+

Acknowledgements
----------------

Shamelessly inspired by Zappa (https://github.com/Miserlou/Zappa)

Platform: UNKNOWN
Classifier: Development Status :: 3 - Alpha
Classifier: Environment :: Console
Classifier: License :: OSI Approved :: MIT License
Classifier: Operating System :: OS Independent
Classifier: Natural Language :: English
Classifier: Programming Language :: Python
Classifier: Programming Language :: Python :: 3.5

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Filename, size & hash SHA256 hash help File type Python version Upload date
bokchoi-0.4.1.tar.gz (19.5 kB) Copy SHA256 hash SHA256 Source None

Supported by

Elastic Elastic Search Pingdom Pingdom Monitoring Google Google BigQuery Sentry Sentry Error logging AWS AWS Cloud computing DataDog DataDog Monitoring Fastly Fastly CDN SignalFx SignalFx Supporter DigiCert DigiCert EV certificate StatusPage StatusPage Status page