Skip to main content

TF-Slim models (Guild AI)

Project description

gpkg.slim.models
################

*TF-Slim models (Guild AI)*

Models
######

images
======

*Generic images dataset*

Operations
^^^^^^^^^^

prepare
-------

*Prepare images for training*

Flags
`````

**images**
*Directory containing images to prepare (required)*

**random-seed**
*Seed used for train/validation split (randomly generated)*

**val-split**
*Percentage of images reserved for validation (30)*


inception
=========

*TF-Slim Inception v1 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


inception-resnet-v2
===================

*TF-Slim Inception ResNet v2 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


inception-v2
============

*TF-Slim Inception v2 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


inception-v3
============

*TF-Slim Inception v3 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


inception-v4
============

*TF-Slim Inception v4 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


mobilenet
=========

*TF-Slim Mobilenet v1 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


mobilenet-v2-1.4
================

*TF-Slim Mobilenet v2 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


nasnet-large
============

*TF-Slim NASNet large classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


nasnet-mobile
=============

*TF-Slim NASNet mobile classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


pnasnet-large
=============

*TF-Slim PNASNet classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


pnasnet-mobile
==============

*TF-Slim PNASNet mobile classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


resnet-101
==========

*TF-Slim ResNet v1 101 layer classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


resnet-152
==========

*TF-Slim ResNet v1 152 layer classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


resnet-50
=========

*TF-Slim ResNet v1 50 layer classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


resnet-v2-101
=============

*TF-Slim ResNet v2 101 layer classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


resnet-v2-152
=============

*TF-Slim ResNet v2 152 layer classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


resnet-v2-50
============

*TF-Slim ResNet v2 50 layer classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


vgg-16
======

*TF-Slim VGG 16 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*


vgg-19
======

*TF-Slim VGG 19 classifier*

Operations
^^^^^^^^^^

evaluate
--------

*Evaluate a trained model*

Flags
`````

**batch-size**
*Number of examples in each evaluated batch (100)*

**eval-batches**
*Number of batches to evaluate (all available)*

**step**
*Checkpoint step to evaluate (latest checkpoint)*

export-and-freeze
-----------------

*Export an inference graph with checkpoint weights*

Flags
`````

**step**
*Checkpoint step to use for the frozen graph (latest checkpoint)*

finetune
--------

*Finetune a trained model*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.0001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

label
-----

*Classify an image using a trained model*

Flags
`````

**image**
*Path to image to classify (required)*

tflite
------

*Generate a TFLite file from a frozen graph*

Flags
`````

**output-format**
*TF Lite output format (tflite)

Choices:
tflite
graphviz_dot

*

**quantized**
*Whether or not output arrays are quantized (no)

Choices:
yes
no

*

**quantized-inputs**
*Whether or not input arrays are quantized (no)

Choices:
yes
no

*

train
-----

*Train model from scratch*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*

transfer-learn
--------------

*Train model using transfer learning*

Flags
`````

**auto-scale**
*Adjust applicable flags for multi-GPU systems (yes)

Set to 'no' to disable any flag value adjustments.

When this value is 'yes' (the default) the following flags are adjusted on
multi-GPU systems:

- clones

- learning-rate

`clones` is set to the number of available GPUs.

`learning-rate` is adjusted by multiplying its specified value by the
number of GPUs.

Flags are not adjusted on single GPU or CPU only systems.

*

**batch-size**
*Number of examples in each training batch (32)*

**clones**
*Number of model clones (calculated)

This value is automatically set to the number of available GPUs if `auto-
scale` is 'yes'.

When `auto-scale` is 'no' this value can be increased from 1 to train the
model in parallel on multiple GPUs.

*

**learning-rate**
*Initial learning rate (0.001)*

**learning-rate-decay-epochs**
*Number of epochs after which learning rate decays (2.0)*

**learning-rate-decay-factor**
*Learning rate decay factor (0.94)*

**learning-rate-decay-type**
*Method used to decay the learning rate (exponential)

Choices:
exponential
fixed
polynomial

*

**learning-rate-end**
*Minimal learning rate used by polynomial learning rate decay (0.0001)*

**log-save-seconds**
*Frequency of log summary saves in seconds (60)*

**log-steps**
*Frequency of summary logs in steps (100)*

**model-save-seconds**
*Frequency of model saves (checkpoints) in seconds (600)*

**optimizer**
*Optimizer used to train (rmsprop)

Choices:
adadelta
adagrad
adam
ftrl
momentum
rmsprop
sgd

*

**preprocessing**
*Preprocessing to use (default for model)*

**preprocessors**
*Number of preprocessing threads (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize the
preprocessor thread count for the system.

*

**readers**
*Number of parallel data readers (calculated)

This value is automatically set to logical CPU count / 2 if `auto-scale`
is 'yes'.

When `auto-scale` is 'no' this value can be set to optimize data reader
performance for the system.

*

**train-steps**
*Number of steps to train (train indefinitely)*

**weight-decay**
*Decay on the model weights (4e-05)*



Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

gpkg.slim.models-0.5.1-py2.py3-none-any.whl (7.6 kB view details)

Uploaded Python 2 Python 3

File details

Details for the file gpkg.slim.models-0.5.1-py2.py3-none-any.whl.

File metadata

  • Download URL: gpkg.slim.models-0.5.1-py2.py3-none-any.whl
  • Upload date:
  • Size: 7.6 kB
  • Tags: Python 2, Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.9.1 pkginfo/1.4.2 requests/2.18.4 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.19.6 CPython/2.7.12

File hashes

Hashes for gpkg.slim.models-0.5.1-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 63c29c501fa0f6220175599a79df3bd06ff21bc27aba6c096cac49459afc6b3e
MD5 0480cf258d67c3ab3cc38e53a6bcb16b
BLAKE2b-256 6e17917a705fd314d6ab01c23ae4d3204ecb41583a72866f641073490a37eb68

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page