VGSLify is a Python toolkit designed for rapid prototyping and seamless conversion between TensorFlow models and the Variable-size Graph Specification Language (VGSL). Drawing inspiration from Tesseract's VGSL specs, VGSLify introduces enhancements and provides a streamlined approach to define, train, and interpret deep learning models using VGSL.
Project description
VGSLify: Variable-size Graph Specification for TensorFlow & PyTorch
VGSLify is a powerful toolkit designed to simplify the process of defining, training, and interpreting deep learning models using the Variable-size Graph Specification Language (VGSL). Inspired by Tesseract's VGSL specs, VGSLify introduces a set of enhancements and provides a streamlined interface for both TensorFlow and, in future releases, PyTorch.
Table of Contents
- Installation
- How VGSL Works
- Quick Start
- Supported Layers and Their Specifications
- Future Work
- Contributing
- License
- Acknowledgements
Installation
Installing VGSLify is simple and straightforward. Follow the steps below to get started:
1. Prerequisites
Before installing VGSLify, ensure you have Python 3.8 or newer installed. You can check your Python version with the following command:
python --version
If you don't have Python installed, download it from python.org or use your system's package manager.
2. Using pip
The easiest way to install VGSLify is via pip
. Run the following command to install the latest version of VGSLify:
pip install vgslify
3. From Source
If you want to install the development version or a modified version of VGSLify, you can do so from source. First, clone the repository or download the source code. Then navigate to the directory and run:
pip install .
4. Installing a backend
VGSLify allows you to choose your preferred deep learning framework. You must install the backend separately according to your needs:
-
TensorFlow: To use VGSLify with TensorFlow, install TensorFlow using the following command:
pip install tensorflow
-
PyTorch: For PyTorch support, install PyTorch using the following command:
pip install torch
You can install both backends if you want to switch between TensorFlow and PyTorch.
5. Verifying Installation
After installation, you can verify that VGSLify was installed correctly by importing it in Python:
python -c "import vgslify; print(vgslify.__version__)"
This should print the installed version of VGSLify without any errors.
Remember to periodically update VGSLify to benefit from new features and bug fixes. You can update it using pip
:
pip install --upgrade vgslify
How VGSL works
VGSL operates through short definition strings. For instance:
None,None,64,1 Cr3,3,32 Mp2,2,2,2 Cr3,3,64 Mp2,2,2,2 Rc Fc64 D20 Lrs128 D20 Lrs64 D20 O1s92
In this example, the string defines a neural network with input layers, convolutional layers, pooling, reshaping, fully connected layers, LSTM and output layers. Each segment of the string corresponds to a specific layer or operation in the neural network. Moreover, VGSL provides the flexibility to specify the type of activation function for certain layers, enhancing customization.
Quick Start
Using VGSLify, you can rapidly prototype TensorFlow models using the Variable-size Graph Specification Language (VGSL). Here's a simple example to get you started:
Generating a Model with VGSLify
-
Import the Necessary Module:
First, import the
VGSLModelGenerator
class from VGSLify's core module.from vgslify.generator import VGSLModelGenerator
-
Initialize the Model Generator:
VGSLify is a "BYOB" (Bring Your Own Backend) package, meaning you can choose the backend that suits your needs. Options include:
- "tensorflow": Use TensorFlow for model generation.
- "torch": Use PyTorch as the backend (currently under development).
- "auto": VGSLify will automatically detect the available backend and use TensorFlow by default if both are available.
For example, to explicitly use TensorFlow as the backend:
vgsl_gn = VGSLModelGenerator(backend="tensorflow")
Alternatively, let VGSLify automatically detect the backend:
vgsl_gn = VGSLModelGenerator(backend="auto")
or
vgsl_gn = VGSLModelGenerator()
-
Build and View the Model:
After initializing the model generator, you can pass the VGSL specification string to the
generate_model()
method to build the model. Here's an example where we create a TensorFlow model with a convolutional layer, max-pooling layer, and an output softmax layer:vgsl_spec = "None,64,None,1 Cr3,3,32 Mp2,2,2,2 O1s92" model = vgsl_gn.generate_model(vgsl_spec) model.summary()
This example demonstrates the flexibility and simplicity of creating a TensorFlow model using VGSLify. The VGSL spec string allows you to define complex architectures with ease, and the VGSLModelGenerator
handles the construction of the model. You can use the same generator instance to create multiple models by passing different VGSL spec strings.
For more details on the VGSL spec syntax and supported layers, refer to the "Supported Layers and Their Specifications" section.
Creating Individual Layers with VGSLify
In addition to creating complete models, VGSLify allows you to generate individual layers from a VGSL specification string. This is particularly useful when you want to integrate a VGSL-defined layer into an existing model or when you wish to experiment with individual components without having to construct a full model.
-
Import the Necessary Module:
Just like before, you'll need to import the
VGSLModelGenerator
class.from vgslify.generator import VGSLModelGenerator
-
Initialize the Model Generator:
Like when building models, you first initialize the model generator and select your backend. In this example, we'll use TensorFlow:
vgsl_gn = VGSLModelGenerator(backend="tensorflow")
-
Generate an Individual Layer:
You can now use the
construct_layer()
method of VGSLModelGenerator to create individual layers directly from a VGSL spec string.Here’s an example where we create a convolutional layer using the VGSL spec:
vgsl_spec_for_conv2d = "Cr3,3,64" conv2d_layer = vgsl_gn.construct_layer(vgsl_spec_for_conv2d)
These layers can be added to an existing TensorFlow model just like any other
tf.keras.layers.Layer
. -
Generate the History of Layers:
If you want to inspect the sequence of layers from a VGSL specification without constructing a full model, you can use the
generate_history()
method. This method parses the VGSL spec and generates the layers in sequence, allowing you to experiment with or debug the architecture.Here’s how you can use it:
vgsl_spec = "None,64,None,1 Cr3,3,32 Mp2,2,2,2 O1s92" history = vgsl_gn.generate_history(vgsl_spec) for layer in history: print(layer)
This will print each layer generated by the VGSL spec, giving you insight into the layer sequence without building or chaining them together.
Converting TensorFlow Models to VGSL [WIP]
Once you have trained a TensorFlow model, you might want to convert it back into a VGSL spec string for various purposes. VGSLify provides an easy way to do this:
-
Load Your TensorFlow Model:
If you've saved your trained Keras model to a file, first load it using TensorFlow's
load_model
method:import tensorflow as tf model = tf.keras.models.load_model("path_to_your_model.keras")
-
Convert to VGSL Spec String:
With your model loaded, use the
model_to_spec
function from VGSLify to convert the model into a VGSL spec string:from vgslify.utils import model_to_spec vgsl_spec_string = model_to_spec(model) print(vgsl_spec_string)
This provides a concise representation of your model's architecture in VGSL format. Please note that while VGSLify aims to support a wide variety of architectures, there might be specific TensorFlow layers or configurations that are not supported. In such cases, a ValueError
will be raised.
Supported Layers and Their Specifications
Overview
Below is a concise table providing a summary of each supported layer:
Layer | Spec | Example | Description |
---|---|---|---|
Input | batch,height,width,depth |
None,64,None,1 |
Input layer with variable batch size & width, and 1 channel depth |
Output | O(2|1|0)(l|s) |
O1s10 |
Dense layer with a 1D sequence, 10 output classes, and softmax activation |
Conv2D | C(s|t|r|l|m),<x>,<y>[,<s_x>,<s_y>],<d> |
Cr3,3,64 |
Conv2D layer with ReLU activation, 3x3 filter size, 1x1 stride, and 64 filters |
Dense (Fully Connected, FC) | F(s|t|r|l|m)<d> |
Fs64 |
Dense layer with softmax activation and 64 units |
LSTM | L(f|r)[s]<n>[,D<rate>][,Rd<rate>] |
Lf64sD25Rd10 |
LSTM cell (forward-only) with 64 units, return sequences, 0.25 dropout, and 0.10 recurrent dropout |
GRU | G(f|r)[s]<n>[,D<rate>][,Rd<rate>] |
Gr64s,D20,Rd15 |
GRU cell (reverse-only) with 64 units, return sequences, 0.20 dropout, and 0.15 recurrent dropout |
Bidirectional | B(g|l)<n>[,D<rate>][,Rd<rate>] |
Bl256,D15,Rd10 |
Bidirectional layer wrapping an LSTM RNN with 256 units, 0.15 dropout, and 0.10 recurrent dropout |
BatchNormalization | Bn |
Bn |
BatchNormalization layer |
MaxPooling2D | Mp<x>,<y>,<s_x>,<s_y> |
Mp2,2,1,1 |
MaxPooling2D layer with 2x2 pool size and 1x1 strides |
AvgPooling2D | Ap<x>,<y>,<s_x>,<s_y> |
Ap2,2,2,2 |
AveragePooling2D layer with 2x2 pool size and 2x2 strides |
Dropout | D<rate> |
D25 |
Dropout layer with a dropout rate of 0.25 |
Reshape | Rc |
Rc |
Reshape layer returns a new (collapsed) tf.Tensor based on the previous layer outputs |
Note: In the specs, the |
symbol indicates options. For example, in O(2 | 1 | 0)(l | s)
, it means the output layer could be O2l
, O1s
, etc. Arguments between the [
and ]
symbol indicate that this is optional. The [s]
in RNN layers activates return_sequences
.
For more detailed information about each layer and its associated VGSL spec, see the following sections:
Layer Details
Input
- Spec:
batch,height,width,depth
- Description: Represents the TensorFlow input layer, based on standard TF tensor dimensions.
- Example:
None,64,None,1
creates atf.keras.layers.Input
with a variable batch size, height of 64, variable width, and a depth of 1 (input channels).
Output
- Spec:
O(2|1|0)(l|s)<n>
- Description: Output layer providing either a 2D vector (heat) map of the input (
2
), a 1D sequence of vector values (1
) or a 0D single vector value (0
) withn
classes. Currently, only a 1D sequence of vector values is supported. - Example:
O1s10
creates a Dense layer with a 1D sequence as output with 10 classes and softmax.
Conv2D
- Spec:
C(s|t|r|l|m)<x>,<y>[,<s_x>,<s_y>],<d>
- Description: Convolutional layer using a
x
,y
window andd
filters. Optionally, the stride window can be set with (s_x
,s_y
). - Examples:
Cr3,3,64
creates a Conv2D layer with a ReLU activation function, a 3x3 filter, 1x1 stride, and 64 filters.Cr3,3,1,3,128
creates a Conv2D layer with a ReLU activation function, a 3x3 filter, 1x3 strides, and 128 filters.
Dense
- Spec:
F(s|t|r|e|l|m)<d>
- Description: Fully-connected layer with
s|t|r|e|l|m
non-linearity andd
units. - Example:
Fs64
creates a FC layer with softmax non-linearity and 64 units.
LSTM
- Spec:
L(f|r)[s]<n>[,D<rate>][,Rd<rate>]]
- Description: LSTM cell running either forward-only (
f
) or reversed-only (r
), withn
units. Optionally, therate
can be set for thedropout
and/or therecurrent_dropout
, whererate
indicates a percentage between 0 and 100. - Example:
Lf64
creates a forward-only LSTM cell with 64 units.
GRU
- Spec:
G(f|r)[s]<n>[,D<rate>][,Rd<rate>]
- Description: GRU cell running either forward-only (
f
) or reversed-only (r
), withn
units. Optionally, therate
can be set for thedropout
and/or therecurrent_dropout
, whererate
indicates a percentage between 0 and 100. - Example:
Gf64
creates a forward-only GRU cell with 64 units.
Bidirectional
- Spec:
B(g|l)<n>[,D<rate>][,Rd<rate>]
- Description: Bidirectional layer wrapping either a LSTM (
l
) or GRU (g
) RNN layer, running in both directions, withn
units. Optionally, therate
can be set for thedropout
and/or therecurrent_dropout
, whererate
indicates a percentage between 0 and 100. - Example:
Bl256
creates a Bidirectional RNN layer using a LSTM Cell with 256 units.
BatchNormalization
- Spec:
Bn
- Description: A technique often used to standardize the inputs to a layer for each mini-batch. Helps stabilize the learning process.
- Example:
Bn
applies a transformation maintaining mean output close to 0 and output standard deviation close to 1.
MaxPooling2D
- Spec:
Mp<x>,<y>,<s_x>,<s_y>
- Description: Downsampling technique using a
x
,y
window. The window is shifted by stridess_x
,s_y
. - Example:
Mp2,2,2,2
creates a MaxPooling2D layer with pool size (2,2) and strides of (2,2).
AvgPooling2D
- Spec:
Ap<x>,<y>,<s_x>,<s_y>
- Description: Downsampling technique using a
x
,y
window. The window is shifted by stridess_x
,s_y
. - Example:
Ap2,2,2,2
creates an AveragePooling2D layer with pool size (2,2) and strides of (2,2).
Dropout
- Spec:
D<rate>
- Description: Regularization layer that sets input units to 0 at a rate of
rate
during training. Used to prevent overfitting. - Example:
D50
creates a Dropout layer with a dropout rate of 0.5 (D
/100).
Reshape
- Spec:
Rc
- Description: Reshapes the output tensor from the previous layer, making it compatible with RNN layers.
- Example:
Rc
applies a specific transformation:layers.Reshape((-1, prev_layer_y * prev_layer_x))
.
Future Work
PyTorch Model Support
While the current version of VGSLify supports TensorFlow models, I recognize the growing popularity and capabilities of PyTorch in the deep learning community. In an upcoming release, I aim to expand VGSLify's capabilities to generate and parse PyTorch models using VGSL spec, providing a unified experience across both major deep learning frameworks.
Custom Layer Support
To make VGSLify even more versatile, I'm working on a feature that will allow users to define and integrate custom layers into the VGSL specification. This enhancement will empower users to seamlessly integrate specialized layers or proprietary architectures into their VGSL-defined models, further bridging the gap between rapid prototyping and production-ready models.
Spec to Code
The Spec to Code feature will generate Python code from a VGSL spec string, allowing users to easily view, customize, and integrate their model architectures into existing codebases. This enhances transparency and flexibility, giving users full control over the generated model, allowing them to optimize and tune it as needed.
Model to Code
The Model to Code feature will convert trained models back into maintainable Python code, allowing easy integration, modification, and deployment. It will support both TensorFlow and PyTorch, generating framework-specific code based on the model’s architecture.
Contributing
I warmly welcome contributions to VGSLify! Whether you're fixing bugs, adding new features, or improving the documentation, your efforts will make VGSLify better for everyone.
How to Contribute:
- Fork the Repository: Start by forking the VGSLify repository.
- Set Up Your Development Environment: Clone your fork to your local machine and set up the development environment.
- Make Your Changes: Implement your changes, improvements, or fixes.
- Submit a Pull Request: Once you're done with your changes, push them to your fork and submit a pull request. Please provide a clear description of the changes and their purpose.
- Create Issues: If you find bugs or want to suggest improvements, please create an issue in the repository.
Please ensure that your contributions adhere to our coding standards and practices. Your efforts will help VGSLify grow and thrive!
License
VGSLify is open-source software and is licensed under the MIT License. This means you can freely use, modify, and distribute it, provided you include the original copyright notice. For more details, you can refer to the LICENSE file in the repository.
Acknowledgements
A special thank you to:
- The creators and contributors of the original VGSL specification, which inspired and laid the groundwork for VGSLify.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file vgslify-0.12.0.tar.gz
.
File metadata
- Download URL: vgslify-0.12.0.tar.gz
- Upload date:
- Size: 25.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.10.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e62c55383ccb8afa3660bd84924cb62d1db4894d58c23556e54c4254dc74c46e |
|
MD5 | 7b4f69f9a0d8bf338f9da8b7c0568b42 |
|
BLAKE2b-256 | 125dde039ef2dcd491517ebe0ee1f489ecc65571b508c930f3c615005fea0caa |
File details
Details for the file vgslify-0.12.0-py3-none-any.whl
.
File metadata
- Download URL: vgslify-0.12.0-py3-none-any.whl
- Upload date:
- Size: 22.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.10.12
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | c2ec91063a40455328447bcd4ef27f0e4cd8da2a98d67837751687be663f3fbf |
|
MD5 | 5587acd278cd329c755aa4f083225364 |
|
BLAKE2b-256 | 77b7e0b7ee0857f6b28f03cabb8225381ba8d5bfee5327b3e426c59d10d7be4e |