Skip to main content
Join the official Python Developers Survey 2018 and win valuable prizes: Start the survey!

Azkaban CLI

Project description

Lightweight command line interface (CLI) for Azkaban:

  • Define jobs from a single python file
  • Build projects and upload to Azkaban from the command line

Integration is meant to be as transparent as possible:

  • No additional folders and files
  • No imposed project structure

Installation

Using pip:

$ pip install azkaban

Quickstart

We first create a file to define our project. Let’s call it jobs.py, although any name would work.

In this example, we add a single job and file:

from azkaban import Job, Project

project = Project('foo')

project.add_file('/path/to/bar.txt', 'bar.txt')
project.add_job('bar', Job({'type': 'command', 'command': 'cat bar.txt'}))

if __name__ == '__main__':
  project.main()

From the command line we can now run python jobs.py --help to view the list of all available options (build, upload, etc.). E.g. the following command will create the archive foo.zip containing all the project’s jobs and dependency files:

$ python jobs.py build foo.zip

More

Aliases

To avoid having to enter the server’s URL on every upload (or hard-coding it into our project’s configuration file, ugh), we can define aliases in ~/.azkabanrc:

[foo]
url = http://url.to.foo.server:port
[bar]
url = http://url.to.bar.server:port

We can now upload directly to each of these URLs with the shorthand:

$ python jobs.py upload -a foo

This has the added benefit that we won’t have to authenticate on every upload. The session ID is cached and reused for later connections.

Job options

There often are options which are common across multiple jobs. For this reason, the Job constructor takes in multiple options dictionaries. The first definition of an option (i.e. earlier in the arguments) will take precedence over later ones.

We can use this to efficiently share default options among jobs, for example:

defaults = {'user.to.proxy': 'boo', 'retries': 0}
jobs = [
  Job({'type': 'noop'}),
  Job({'type': 'noop'}, defaults),
  Job({'type': 'command', 'command': 'ls'}, defaults),
  Job({'type': 'command', 'command': 'ls -l', 'retries': 1}, defaults),
]

All jobs except the first one will have their user.to.proxy property set. Note also that the last job overrides the retries property.

Finally, nested dictionaries can be used to group options efficiently:

# e.g. this job
Job({
  'proxy.user': 'boo',
  'proxy.keytab.location': '/path',
  'param.input': 'foo',
  'param.output': 'bar',
})
# is equivalent to this one
Job({
  'proxy': {'user': 'boo', 'keytab.location': '/path'},
  'param': {'input': 'foo', 'output': 'bar'}
})

Pig jobs

Because pig jobs are so common, a PigJob class is provided which accepts a file path (to the pig script) as first constructor argument, optionally followed by job options. It then automatically sets the job type and adds the corresponding script file to the project.

from azkaban import PigJob

project.add_job('baz', PigJob('/.../baz.pig', {'dependencies': 'bar'}))

Next steps

Any valid python code can go inside the jobs configuration file. This includes using loops to add jobs, subclassing the base Job class to better suit a project’s needs (e.g. by implementing the on_add and on_build handlers), …

Project details


Release history Release notifications

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Filename, size & hash SHA256 hash help File type Python version Upload date
azkaban-0.1.2.tar.gz (6.5 kB) Copy SHA256 hash SHA256 Source None Oct 18, 2013

Supported by

Elastic Elastic Search Pingdom Pingdom Monitoring Google Google BigQuery Sentry Sentry Error logging AWS AWS Cloud computing DataDog DataDog Monitoring Fastly Fastly CDN SignalFx SignalFx Supporter DigiCert DigiCert EV certificate StatusPage StatusPage Status page