No project description provided
Project description
IRIS CLI Package.
Description
Iris is your portal to the TitanML platform. Using Iris, you can launch jobs to run on TitanML servers, run your own models and datasets through our compression algorithms, and explore and download the optimised models from the Titan Store. The backend takes one of the following supported models and uses it to finetune similar, smaller models. The training signals from the teacher model improve the performance of the student model on edge cases, and allow you to use cheaper, more readily-available unlabelled data.
Getting Started
Dependencies
- python >= 3.7
- titanML login
Installing
- using pip
pip install titan-iris
iris API
iris
Usage:
$ iris [OPTIONS] COMMAND [ARGS]...
Options:
--help
: Show this message and exit.
Commands:
delete
: delete objects from the TYTN api.download
: Download the titan-optimized onnx model.get
: Get objects from the TYTN api.infer
: Run inference on a model.login
: Login to iris.logout
: Logout from iris.makesafe
: Convert a non-safetensor model into a...post
: Dispatch a job to the TitanML platformpull
: Pull the titan-optimized server docker image.status
: Get the status of an experimentupload
: Upload an artefact to the TitanML hub.
iris delete
delete objects from the TYTN api.
Usage:
$ iris delete [OPTIONS] [OBJECT]:[experiment|artefact]
Arguments:
[OBJECT]:[experiment|artefact]
: What type of object to delete [default: experiment]
Options:
-i, --id TEXT
: Which object to delete [required]--help
: Show this message and exit.
iris download
Download the titan-optimized onnx model.
Usage:
$ iris download [OPTIONS] IMAGE
Arguments:
IMAGE
: The model to pull. Should be displayed in the TitanML Hub. [required]
iris get
Get objects from the TYTN api.
Usage:
$ iris get [OPTIONS] [OBJECT]:[experiment|artefact]
Arguments:
[OBJECT]:[experiment|artefact]
: What type of object to get [default: experiment]
Options:
-i, --id TEXT
: Which object to get. None, or '' correspond to getting all objects. Evaluated server-side.-q, --query TEXT
: A JMESPath string, to filter the objects returned by the API. Evaluated client-side.-h, --headers TEXT
: Headers to send with the get request. Should be provided as colon separated key value pairs: -h a:b -h c:d -> {a:b, c:d} [default: ]--help
: Show this message and exit.
iris infer
Run inference on a model.
Usage:
$ iris infer [OPTIONS]
Options:
--target TEXT
: The url to run the server on. [default: localhost]-p, --port INTEGER
: The port to run the server on. [default: 8000]-t, --task [sequence_classification|glue|question_answering|token_classification]
: The task to optimize the model for. [required]--use-cpu
: Whether to use the CPU. If False, the GPU will be used. Choose CPU only when the opmitized model is in CPU format(OnnxRuntime). The default will be False. (using TensorRT) [default: False]-t, --text TEXT
: The text to run the server in. In classification tasks, this is the TEXT to be classified. In question answering tasks, this is the QUESTION to be answered. [required]-c, --context TEXT
: The context in question answering tasks. Only used in question answering tasks. [default: ]--help
: Show this message and exit.
iris login
Login to iris.
Usage:
$ iris login [OPTIONS]
iris logout
Logout from iris.
Usage:
$ iris logout [OPTIONS]
iris makesafe
Convert a non-safetensor model into a safetensor model, including for models with shared weights.
Usage:
$ iris makesafe [OPTIONS] [MODEL]
Arguments:
[MODEL]
: The model to convert to safe_tensors [default: model]
iris post
Dispatch a job to the TitanML platform.
Usage:
$ iris post [OPTIONS]
Options:
-m, --model TEXT
: The model to optimize. [required]-d, --dataset TEXT
: The dataset to optimize the model with. [required]-t, --task [sequence_classification|glue|question_answering|token_classification]
: The task to optimize the model for. [required]-n, --name TEXT
: The name to use for this job. Visible in the TitanML Hub. [default: ]-f, --file TEXT
: Load the options from a config file [default: ]-s, --short-run
: Truncates the run after 1 batch and 1 epoch. Will provide bad results, but useful to check that the model and dataset choices are valid. [default: False]-nl, --num-labels INTEGER
: Number of labels. Required for task sequence_classification-tf, --text-fields TEXT
: Text fields. Required for task sequence_classification-hn, --has-negative
: Has negative. Required for question_answering [default: False]-ln, --label-names TEXT
: Names of token labels. Required for task token_classification. Specify as a mapping with no spaces:-ln 0:label1 -ln 1:label2
--help
: Show this message and exit.
iris pull
Pull the titan-optimized server docker image.
Usage:
$ iris pull [OPTIONS] IMAGE
Arguments:
IMAGE
: The image to pull. Should be displayed in the TitanML Hub. [required]
iris status
Get the status of an experiment
Usage:
$ iris status [OPTIONS]
Options:
-i, --id INTEGER
: The id of the experiment to get the status of [required]
iris upload
Upload an artefact to the TitanML Hub.
Usage:
$ iris upload [OPTIONS] SRC [NAME] [DESCRIPTION]
Arguments:
SRC
: The location of the artefact on disk. Should be a folder, containing either a model or a dataset. For more information on the supported formats, see here. [required][NAME]
: The name of the artefact. Displayed in the TitanMl Hub.[DESCRIPTION]
: A short description of the artefact. Displayed in the TitanML Hub.
Options:
--help
: Show this message and exit.
Help
Any advise for common problems or issues.
command to run if program contains helper info
Authors
TitanML
Version History
License
Acknowledgments
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for titan_iris-0.8.14-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | b9592e8aa53829471bb1dc07c4d5523d33d8927af4474a42bf8fef3a87413f0e |
|
MD5 | 0fb738029517be84495b658ec3b7bdb1 |
|
BLAKE2b-256 | 6ba1b76c14ec4d21eec11ea1b81fda2155caac436929a9bf6b18779985e1cffc |