CLI library to automate the onboarding process to IBM Watson OpenScale
Project description
ibm-ai-openscale-cli
IBM Watson Openscale "express path" configuration tool. This tool allows the user to get started quickly with Watson OpenScale.
- If needed, automatically provision a Lite plan instance for IBM Watson OpenScale
- If needed, automatically provision a Lite plan instance for IBM Watson Machine Learning
- Drop and re-create the IBM Watson OpenScale datamart instance and datamart database schema
- Optionally, deploy a sample machine learning model to the WML instance
- Configure the sample model instance to OpenScale, including payload logging, fairness checking, feedback, quality checking, drift checking, business KPI correlation checking, and explainability
- Optionally, store up to 7 days of historical payload, fairness, quality, drift, and business KPI correlation data for the sample model
- Upload new feedback data, generate 100 new live scoring predictions, run fairness, quality, drift, and correlation checks, and generate one explanation
What's new in this release
- Support for business KPI correlation monitoring added to the built-in sample GermanCreditRisk model, including 7 days of correlation history
- Support for WML v4 python client, using the new
--v4
option. Note that this requires some manual intervention:
- manually uninstall the regular watson-machine-learning-client python package (if installed),
- manually install the watson-machine-learning-client-V4 python package,
- and only then install or upgrade ibm-ai-openscale-cli.
- Other bug fixes and stability improvements
Before you begin
- You need an IBM Cloud account.
- Create an IBM Cloud API key
- If you already have a Watson Machine Learning (WML) instance, ensure it's RC-enabled, learn more about this in the migration instructions.
Installation
To install, use pip
or easy_install
:
pip install -U ibm-ai-openscale-cli
or
easy_install -U ibm-ai-openscale-cli
Usage
ibm-ai-openscale-cli --help
usage: ibm-ai-openscale-cli [-h] (-a APIKEY | -i IAM_TOKEN)
[--env {ypprod,ypqa,ypcr,ys1dev,icp}]
[--resource-group RESOURCE_GROUP]
[--postgres POSTGRES] [--icd ICD] [--db2 DB2]
[--wml WML] [--azure-studio AZURE_STUDIO]
[--azure-service AZURE_SERVICE] [--spss SPSS]
[--custom CUSTOM] [--aws AWS]
[--deployment-name DEPLOYMENT_NAME]
[--keep-schema] [--username USERNAME]
[--password PASSWORD] [--url URL]
[--datamart-name DATAMART_NAME]
[--history HISTORY] [--history-only]
[--history-first-day HISTORY_FIRST_DAY]
[--model MODEL] [--list-models]
[--custom-model CUSTOM_MODEL]
[--custom-model-directory CUSTOM_MODEL_DIRECTORY]
[--extend] [--protect-datamart]
[--reset {metrics,monitors,datamart,model,all}]
[--verbose] [--version] [--v4]
[--wml-plan {lite,standard,professional}]
[--openscale-plan {lite,standard}]
[--bkpi][--generate-drift-history]
IBM Watson Openscale "express path" configuration tool. This tool allows the
user to get started quickly with Watson OpenScale: 1) If needed, provision a
Lite plan instance for IBM Watson OpenScale 2) If needed, provision a Lite
plan instance for IBM Watson Machine Learning 3) Drop and re-create the IBM
Watson OpenScale datamart instance and datamart database schema 4) Optionally,
deploy a sample machine learning model to the WML instance 5) Configure the
sample model instance to OpenScale, including payload logging, fairness
checking, feedback, quality checking, drift, business KPI, and explainability
6) Optionally, store up to 7 days of historical payload, fairness, quality,
drift, and business KPI data for the sample model 7) Upload new feedback data,
generate 100 new live scoring predictions, run fairness, quality, drift, and
business KPI checks, and generate one explanation
optional arguments:
-h, --help show this help message and exit
--env {ypprod,ypqa,ypcr,ys1dev,icp}
Environment. Default "ypprod"
--resource-group RESOURCE_GROUP
Resource Group to use. If not specified, then
"default" group is used
--postgres POSTGRES Path to postgres credentials file for the datamart
database. If --postgres, --icd, and --db2 all are not
specified, then the internal Watson OpenScale database
is used
--icd ICD Path to IBM Cloud Database credentials file for the
datamart database
--db2 DB2 Path to IBM DB2 credentials file for the datamart
database
--wml WML Path to IBM WML credentials file
--azure-studio AZURE_STUDIO
Path to Microsoft Azure credentials file for Microsoft
Azure ML Studio
--azure-service AZURE_SERVICE
Path to Microsoft Azure credentials file for Microsoft
Azure ML Service
--spss SPSS Path to SPSS credentials file
--custom CUSTOM Path to Custom Engine credentials file
--aws AWS Path to Amazon Web Services credentials file
--deployment-name DEPLOYMENT_NAME
Name of the existing deployment to use. Required for
Azure ML Studio, SPSS Engine and Custom ML Engine, but
optional for Watson Machine Learning. Required for
custom models
--keep-schema Use pre-existing datamart schema, only dropping all
tables. If not specified, datamart schema is dropped
and re-created
--username USERNAME ICP username. Required if "icp" environment is chosen,
not required if --iam-token is specified
--password PASSWORD ICP password. Required if "icp" environment is chosen,
not required if --iam-token is specified
--url URL ICP url. Required if "icp" environment is chosen
--datamart-name DATAMART_NAME
Specify data mart name and database schema, default is
the datamart database connection username. For
internal database, the default is "wosfastpath"
--history HISTORY Days of history to preload. Default is 7
--history-only Store history only for existing deployment and
datamart. Requires --extend and --deployment-name also
be specified
--history-first-day HISTORY_FIRST_DAY
Starting day for history. Default is 0
--model MODEL Sample model to set up with Watson OpenScale (default
"GermanCreditRiskModel")
--list-models Lists all available models. If a ML engine is
specified, then modesl specific to that engine are
listed
--custom-model CUSTOM_MODEL
Name of custom model to set up with Watson OpenScale.
If specified, overrides the value set by --model. Also
requires that --custom-model-directory
--custom-model-directory CUSTOM_MODEL_DIRECTORY
Directory with model configuration and metadata files.
Also requires that --custom-model be specified
--extend Extend existing datamart, instead of deleting and
recreating it
--protect-datamart If specified, the setup will exit if an existing
datamart setup is found
--reset {metrics,monitors,datamart,model,all}
Reset existing datamart and/or sample models then exit
--verbose verbose flag
--wml-plan {lite,standard,professional}
If no WML instance exists, then provision one with the
specified plan. Default is "lite", other plans are
paid plans
--openscale-plan {lite,standard}
If no OpenScale instance exists, then provision one
with the specified plan. Default is "lite", other
plans are paid plans
--version show program's version number and exit
--bkpi Enable BKPI support.
--v4 Enable support for WML v4 python client
--generate-drift-history
Generate drift history with live execution instead of
loading from pre-generated history. Only needed for
backward compatiblity for GermanCreditRiskModel in
CP4D v2.1.0.2 (August 2019 GA)
required arguments (only one needed):
-a APIKEY, --apikey APIKEY
IBM Cloud platform user APIKey. If "--env icp" is also
specified, APIKey value is not used.
-i IAM_TOKEN, --iam-token IAM_TOKEN
IBM Cloud authentication IAM token, or IBM Cloud
private authentication IAM token. Format can be
(--iam-token "Bearer <token>") or (--iam-token
<token>)
Examples
In this example, if a WML instance already exists it is used, but if not a new Lite plan instance is provisioned and used. If an OpenScale instance exists, its datamart is dropped and recreated along with its datamart internal database schema. Otherwise, a Lite plan OpenScale instance is provisioned. The GermanCreditRiskModel is stored and deployed in WML, configured to OpenScale, and 7 days' historical data stored. Then new feedback data is uploaded, 100 new live scoring predictions are made, followed by fairness, quality, drift, and business KPI correlation checks, and one explanation.
export APIKEY=<IBM_CLOUD_API_KEY>
ibm-ai-openscale-cli --apikey $APIKEY
In this example, assume the user already has provisioned instances of WML, OpenScale, IBM Cloud Database for Postgres (ICD), and has selected a schema for the OpenScale datamart database. The OpenScale datamart is dropped and recreated, and the datamart's database schema is dropped and recreated. An already-deployed instance of the DrugSelectionModel is configured to OpenScale, and 7 days' historical data stored, followed by new feedback data upload, 100 new scores, fairness, quality, drift, and business KPI correlation checks, and one explanation.
export APIKEY=<IBM_CLOUD_API_KEY>
export WML=<path to WML instance credentials JSON file>
export ICD=<path to ICD instance credentials JSON file>
export SCHEMA=<ICD database schema name>
ibm-ai-openscale-cli --apikey $APIKEY --wml $WML --model DrugSelectionModel --deployment-name DrugSelectionModelDeployment --icd $ICD --datamart-name $SCHEMA
In this example, assume the user already has provisioned an Entry plan instance of IBM DB2 Warehouse on Cloud. The OpenScale datamart's tables within the user's existing DB2 schema are dropped and recreated. The GermanCreditRiskModel is stored and deployed in WML, configured to OpenScale, and 7 days' historical data stored, followed by new feedback data upload, 100 new scores, fairness, quality, drift, and business KPI correlation checks, and one explanation.
export APIKEY=<IBM_CLOUD_API_KEY>
export DB2=<path to DB2 instance credentials JSON file>
export SCHEMA=<user's DB2 database schema>
ibm-ai-openscale-cli --apikey $APIKEY --db2 $DB2 --datamart-name $SCHEMA --keep-schema
In this example, assume the user has their own custom model named MyBusinessModel stored in WML and deployed as MyBusinessModelDeployment.
Also assume they already have a provisioned instance of OpenScale which has not yet been configured.
In the custom model directory, the user has provided a configuration.json
file with the required model configuration details.
The OpenScale datamart and datamart database schema are created, and the MyBusinessModelDeployment is configured to OpenScale.
Then new feedback data is uploaded (if provided), 100 new scoring requests are made to the model, followed by fairness and quality checks (if configured), and one explanation.
export APIKEY=<IBM_CLOUD_API_KEY>
export WML=<path to WML instance credentials JSON file>
export MODELPATH=<path to custom model directory>
ibm-ai-openscale-cli --apikey $APIKEY --wml $WML --custom-model MyBusinessModel --deployment-name MyBusinessModelDeployment --custom-model-directory $MODELPATH
FAQ
Q: What is the GermanCreditRiskModel sample model?
A. The GermanCreditRiskModel sample model is taken from the "Watson Studio, Watson Machine Learning and Watson OpenScale samples" GitHub repo, specifically the IBM Watson OpenScale tutorials. When you run ibm-ai-openscale-cli to deploy and configure the GermanCreditRiskModel, the result will be as if you had run the tutorial notebook appropriate for your machine learning engine.
Q: What are the formats for the credentials files?
A: Each credential file has its own format:
Postgres
{
"uri": "postgres://<USERNAME>:<PASSWORD>@<HOSTNAME>:<PORT>/<DB>"
}
IBM Cloud Database for Postgres(ICD)
- Copy the Service Credentials from your ICD service instance in IBM Cloud
DB2
{
"username": "<USERNAME>",
"password": "<PASSWORD>",
"hostname": "<HOSTNAME>",
"port": "<PORT>",
"db": "<DB>"
}
IBM Watson Machine Learning (WML)
- Copy the Service Credentials from your WML service instance in IBM Cloud
Microsoft Azure
{
"client_id": "<CLIENT_ID>",
"client_secret": "<CLIENT_SECRET",
"tenant": "<TENANT>",
"subscription_id": "<SUBSCRIPTION_ID>"
}
SPSS
{
"username": "<USERNAME>",
"password": "<PASSWORD",
"url": "<URL>"
}
Custom engine
{
"url": "<URL>"
}
Amazon Web Services Sagemaker (AWS)
{
"access_key_id": "<ACCESS_KEY_ID>",
"secret_access_key": "<SECRET_ACCESS_KEY>",
"region": "<REGION>"
}
Q: How do the reset options work?
A: The reset options each affect a different level of data in the datamart:
--reset metrics
: Clean up the payload logging table, monitoring history tables etc, so that it restores the system to a fresh state with datamart configured, model deployments added, all monitors configured, but no actual metrics in the system yet. The system is ready to go. Not supported for Watson OpenScale internal databases.--reset monitors
: Remove all configured monitors and corresponding metrics and history, but leave the actual model deployments (if any) in the datamart. User can proceed to configure the monitors via user interface, API, or ibm-ai-openscale-cli.--reset datamart
: "Factory reset" the datamart to a fresh state as if there was not any configuration.--reset model
: Delete the sample models and deployments from WML. Not yet supported for non-WML engines. Does not affect the datamart.--reset all
: Reset both the datamart and sample models.
Q: Can I use SSL for connecting to the datamart DB2 database?
A: Yes. The below options can be used for connecting to a DB2 with SSL:
- DB2 Warehouse on Cloud databases automatically support SSL, using the VCAP json file generated on the "Service Credentials" page.
- For on-prem or ICP4D DB2 databases:
- You can specify the path on the local client machine to a copy of the DB2 server's SSL certificate "arm" file, using an "ssldsn" connection string in the VCAP json file:
{ "hostname": "<ipaddr>", "username": "<uid>", "password": "<pw>", "port": 50000, "db": "<dbname>", "ssldsn": "DATABASE=<dbname>;HOSTNAME=<ipaddr>;PORT=50001;PROTOCOL=TCPIP;UID=<uid>;PWD=<pw>;Security=ssl;SSLServerCertificate=/path_on_local_client_machine_to/db2server_instance.arm;" }
- You can specify the base64-encoded certificate as the
certificate_base64
attribute directly in the credentials along with assl
attribute set to true, as below:
{ "hostname": "<ipaddr>", "username": "<uid>", "password": "<pw>", "port": 50000, "db": "<dbname>", "ssl": true, "certificate_base64":"Base64 encoded SSL certificate" }
If SSL connections are not needed, or not configured on the DB2 server, you can remove the "ssldsn" tag and ibm-ai-openscale-cli will use the non-SSL "dsn" tag instead. If the VCAP has both dsn and ssldsn tags, ibm-ai-openscale-cli will use "ssldsn" tag to create an SSL connection.
Q: What are the contents of a custom model directory?
A: These files are used to configure a custom model to IBM Watson OpenScale:
Required
configuration.json
: the model configuration details
Optional
model_content.gzip
: exported model file from WML, to be loaded and deployed into WML if--deployment-name
is not specifiedmodel_meta.json
: exported model metadata from WML (required if model gzip is provided)pipeline_content.gzip
: exported model pipeline file from WML, to be loaded and deployed into WML if--deployment-name
is not specifiedpipeline_meta.json
: exported model pipeline metadata from WML (required if pipeline gzip is provided)drift_model.gzip
: exported model file from WML for a trained Drift model (required if drift configuration provided in configuration.json)
Syntax of configuration.json
A JSON file that specifies the OpenScale configuration for the model. The key components are:
asset_metadata
(required): top-level model specification elementstraining_data_reference
(required): reference to the model training data csv in COStraining_data_type
(optional): required if there are any numeric-valued model featuresquality_configuration
(optional): if applicable for the modelfairness_configuration
(optional): if applicable for the modeldrift_configuration
(optional): if applicable for the model
Valid values for parameters in asset_metadata
:
problem_type
:REGRESSION
,BINARY_CLASSIFICATION
,MULTICLASS_CLASSIFICATION
input_data_type
:STRUCTURED
Here is an example:
{
"asset_metadata": {
"problem_type": "BINARY_CLASSIFICATION",
"input_data_type": "STRUCTURED",
"label_column": "Risk",
"prediction_column": "Scored Labels",
"probability_column": "Scored Probabilities",
"categorical_columns": [ "CheckingStatus" ],
"feature_columns": [ "CheckingStatus", "LoanDuration", "Age" ]
},
"training_data_reference": {
"credentials" : {<IBM Cloud COS credentials>},
"path" : "<path within COS to training data csv file (bucket name + / + filename)>",
"firstlineheader": "True"
},
"training_data_type": { "LoanDuration": "int", "Age": "int" },
"quality_configuration": { "threshold": 0.95, "min_records": 40 },
"fairness_configuration": {
"features": [
{
"feature": "Age",
"majority": [[ 26, 75 ]],
"minority": [[ 18, 25 ]],
"threshold": 0.98
}
],
"favourable_classes": [ "No Risk" ],
"unfavourable_classes": [ "Risk" ],
"min_records": 100
},
"drift_configuration": {
"threshold": 0.15,
"min_records": 100
}
}
Syntax of training_data.csv
A CSV file of the data used to train the model.
This data is also used by live scoring requests to the model using the range of actual values for each feature from the training data.
A header row is required, with column names that match the model's feature names.
Any column with numeric values must be included in the training_data_type
specification in the configuration.json
.
A typical example:
CheckingStatus,LoanDuration,Age,Risk
no_checking,28,30,Risk
0_to_200,28,27,No Risk
. . .
Python version
Tested on Python 3.5, 3.6, and 3.7.
Contributing
See CONTRIBUTING.md.
License
This library is licensed under the Apache 2.0 license.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for ibm-ai-openscale-cli-0.2.391.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6c0128dd845a99d8cdfff54b2e98ab49d455d35d2bd38e1d8c786e65ea7d82ca |
|
MD5 | ba0cfb6a883966131ff176dcd7984a6c |
|
BLAKE2b-256 | 4df7bc321985f291261c2ebdac39eab0ed0f379370a161c22d94dbb6e8bacfa8 |
Hashes for ibm_ai_openscale_cli-0.2.391-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 292bcf0e5b60b38e1218b5cd6ef619f43ed27d3c6dba1f2d7fdb45cd4622dc11 |
|
MD5 | d1ea3d5e6cd07dd25d75962bfd568299 |
|
BLAKE2b-256 | 6f9511c29a413b35154c485474ae7ccdcc6ff52518508203f3841dabf53ffff3 |