Skip to main content

No project description provided

Project description

IRIS CLI Package.

Description

Iris is your portal to the Titan platform. Using Iris, you can launch jobs to run on Titan servers, run your own models and datasets through our compression algorithms, and explore and download the optimised models from the Titan platform. The backend takes one of the following supported models and uses it to finetune similar, smaller models. The training signals from the teacher model improve the performance of the student model on edge cases, and allow you to use cheaper, more readily-available unlabelled data.

Getting Started

Dependencies

  • python >= 3.7
  • titanML login

Installing

  • using pip
pip install titan-iris

iris API

iris

Usage:

$ iris [OPTIONS] COMMAND [ARGS]...

Options:

  • --help: Show this message and exit.

Commands:

  • delete: delete objects from the TYTN api.
  • download: Download the titan-optimized onnx model.
  • get: Get objects from the TYTN api.
  • infer: Run inference on a model.
  • login: Login to the iris client.
  • logout: Logout from the iris client.
  • makesafe: Convert a non-safetensor model into a...
  • post: Dispatch a job to the titan platform
  • pull: Pull the titan-optimized server docker image.
  • status: Get the status of an experiment
  • upload: Upload an artefact to the titan hub.

iris delete

delete objects from the TYTN api.

Usage:

$ iris delete [OPTIONS] [OBJECT]:[experiment|artefact]

Arguments:

  • [OBJECT]:[experiment|artefact]: What type of object to delete [default: experiment]

Options:

  • -i, --id TEXT: Which object to delete [required]
  • --help: Show this message and exit.

iris download

Download the titan-optimized onnx model.

Usage:

$ iris download [OPTIONS] IMAGE

Arguments:

  • IMAGE: The model to pull. Should be displayed in the titan web interface. [required]

iris get

Get objects from the TYTN api.

Usage:

$ iris get [OPTIONS] [OBJECT]:[experiment|artefact]

Arguments:

  • [OBJECT]:[experiment|artefact]: What type of object to get [default: experiment]

Options:

  • -i, --id TEXT: Which object to get. None, or '' correspond to getting all objects. Evaluated server-side.
  • -q, --query TEXT: A JMESPath string, to filter the objects returned by the API. Evaluated client-side.
  • -h, --headers TEXT: Headers to send with the get request. Should be provided as colon separated key value pairs: -h a:b -h c:d -> {a:b, c:d} [default: ]
  • --help: Show this message and exit.

iris infer

Run inference on a model.

Usage:

$ iris infer [OPTIONS]

Options:

  • --target TEXT: The url to run the server on. [default: localhost]
  • -p, --port INTEGER: The port to run the server on. [default: 8000]
  • -t, --task [sequence_classification|glue|question_answering|token_classification]: The task to optimize the model for. [required]
  • --use-cpu: Whether to use the CPU. If False, the GPU will be used. Choose CPU only when the opmitized model is in CPU format(OnnxRuntime). The default will be False. (using TensorRT) [default: False]
  • -t, --text TEXT: The text to run the server in. In classification tasks, this is the TEXT to be classified. In question answering tasks, this is the QUESTION to be answered. [required]
  • -c, --context TEXT: The context in question answering tasks. Only used in question answering tasks. [default: ]
  • --help: Show this message and exit.

iris login

Login to the iris client.

Usage:

$ iris login [OPTIONS]

iris logout

Logout from the iris client.

Usage:

$ iris logout [OPTIONS]

iris makesafe

Convert a non-safetensor model into a safetensor model, including for models with shared weights.

Usage:

$ iris makesafe [OPTIONS] [MODEL]

Arguments:

  • [MODEL]: The model to convert to safe_tensors [default: model]

iris post

Dispatch a job to the titan platform

Usage:

$ iris post [OPTIONS]

Options:

  • -m, --model TEXT: The model to optimize. [required]
  • -d, --dataset TEXT: The dataset to optimize the model with. [required]
  • -t, --task [sequence_classification|glue|question_answering|token_classification]: The task to optimize the model for. [required]
  • -n, --name TEXT: The name to use for this job. Visible in the titan web interface. [default: ]
  • -f, --file TEXT: Load the options from a config file [default: ]
  • -s, --short-run: Truncates the run after 1 batch and 1 epoch. Will provide bad results, but useful to check that the model and dataset choices are valid. [default: False]
  • -nl, --num-labels INTEGER: Number of labels. Required for task sequence_classification
  • -tf, --text-fields TEXT: Text fields. Required for task sequence_classification
  • -hn, --has-negative: Has negative. Required for question_answering [default: False]
  • -ln, --label-names TEXT: Names of token labels. Required for task token_classification. Specify as a mapping with no spaces: -ln 0:label1 -ln 1:label2
  • --help: Show this message and exit.

iris pull

Pull the titan-optimized server docker image.

Usage:

$ iris pull [OPTIONS] IMAGE

Arguments:

  • IMAGE: The image to pull. Should be displayed in the titan web interface. [required]

iris status

Get the status of an experiment

Usage:

$ iris status [OPTIONS]

Options:

  • -i, --id INTEGER: The id of the experiment to get the status of [required]

iris upload

Upload an artefact to the titan hub.

Usage:

$ iris upload [OPTIONS] SRC [NAME] [DESCRIPTION]

Arguments:

  • SRC: The location of the artefact on disk. Should be a folder, containing either a model or a dataset. For more information on the supported formats, see here. [required]
  • [NAME]: The name of the artefact. Displayed in the titan web interface.
  • [DESCRIPTION]: A short description of the artefact. Displayed in the titan web interface.

Options:

  • --help: Show this message and exit.

Help

Any advise for common problems or issues.

command to run if program contains helper info

Authors

TitanML

Version History

License

Acknowledgments

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

titan-iris-0.3.26a2.tar.gz (27.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

titan_iris-0.3.26a2-py3-none-any.whl (27.0 kB view details)

Uploaded Python 3

File details

Details for the file titan-iris-0.3.26a2.tar.gz.

File metadata

  • Download URL: titan-iris-0.3.26a2.tar.gz
  • Upload date:
  • Size: 27.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.16

File hashes

Hashes for titan-iris-0.3.26a2.tar.gz
Algorithm Hash digest
SHA256 8016efb75d0cf89ad3dbdf4e80cd416b1a957edb9557bd8752327820c07f4d4f
MD5 d9b682f808080f6ba23a0fb70a4d767d
BLAKE2b-256 515d7278015d5684383506fbcff9681a93b049969fdf5c9d170293ef902d955c

See more details on using hashes here.

File details

Details for the file titan_iris-0.3.26a2-py3-none-any.whl.

File metadata

  • Download URL: titan_iris-0.3.26a2-py3-none-any.whl
  • Upload date:
  • Size: 27.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.9.16

File hashes

Hashes for titan_iris-0.3.26a2-py3-none-any.whl
Algorithm Hash digest
SHA256 2908f821dc94f04e6d87061d30aa7edf51c9e6cba00316030e66bba68b0a1eb5
MD5 e64372501caf9f93369edebb04ab79b0
BLAKE2b-256 a8dc0fa6069f9b531dec8c5838501c1824a2889cb1cacfbae336c54b98391018

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page