Generate portable C inference code from Keras or PyTorch model
Project description
Copyright 2021 (c) Pierre-Emmanuel Novac penovac@unice.fr Université Côte d'Azur, CNRS, LEAT. All rights reserved.
Qualia-CodeGen-Core
Converts a pre-trained Keras .h5 or PyTorch model to C code for inference.
Generated C code uses channels_last
data format.
Supported layers
- Activation: ReLU (combined to previous Conv1D, Dense, MaxPooling1D, AveragePooling1D AddV2), Softmax
- Conv1D: optional bias, valid padding only
- Dense: optional bias
- MaxPooling1D: valid padding only
- AveragePooling1D: valid padding only
- Flatten: implies reordering next layer's kernel for data format conversion
- ZeroPadding1D: combined with next Conv1D
- AddV2
Dependencies
python >= 3.9
Python:
jinja2
numpy
Keras
Python:
tensorflow >= 2.6.0
keras >= 2.6.0
PyTorch
Python:
torch >= 1.8.0
Installation
pip install -e .
Usage
Generate C code from Keras .h5
qualia_codegen <model.h5> <output directory>
Use in your C code
Include the model: (can also be built as a separate object)
#include "model.h"
Allocate inputs
and outputs
arrays with correct dimensions. Remember that inputs
must have channels_last
data format.
Call it in your C code:
cnn(inputs, outputs);
Add the source file model.c
to your build system. It includes all the other source files for layers, don't add these to the build system.
Examples
See the src/qualia_codegen_core/examples/Linux
directory for a demo console application to evaluate model accuracy.
src/qualia_codegen_core/examples/qualia_codegen-NucleoL476RG
contains an STM32CubeIDE project for the Nucleo-L476RG board that's currently broken due to some recent changes
Documentation
Nothing much…
Source tree
src/qualia_codegen_core/Allocator.py
: manages activation buffer allocation. Tries to group all buffers into one, except when they cannot be overwritten (dependencies).
src/qualia_codegen_core/Converter.py
: the actual conversion code, parses a Keras model and use the template file associated to each layer to generate C code. When weights have to be written, they are optionally quantized to fixed-point by setting the appropriate parameters of Converter
constructor (see its definition)
src/qualia_codegen_core/Validator.py
: work in progress, should contain functions to check if a model can be successfully converted. For now only check activation function.
src/qualia_codegen_core/assets/
: contains the templates to generate C inference code
src/qualia_codegen_core/assets/layers/
: contains the implementation of the various supported layers
src/qualia_codegen_core/assets/layers/weights
: contains the support for the trainable layers weights
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file qualia_codegen_core-2.0.0.tar.gz
.
File metadata
- Download URL: qualia_codegen_core-2.0.0.tar.gz
- Upload date:
- Size: 37.7 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: pdm/2.10.3 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 84c8f4a326c576454cf6d486e90b7a688ed034fd7e55f247f2da81068a8d0101 |
|
MD5 | 00bd5b91b4ab6fe1659a44c6f52f8a7b |
|
BLAKE2b-256 | 1b5e4a99cd4d1617b5e1a362407cec760e76f8d34fee06efe1bfac5464631211 |
File details
Details for the file qualia_codegen_core-2.0.0-py3-none-any.whl
.
File metadata
- Download URL: qualia_codegen_core-2.0.0-py3-none-any.whl
- Upload date:
- Size: 39.1 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: pdm/2.10.3 CPython/3.11.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | c17347e3cd5a1683dd532680afbf7739fe111baaf01ca1a5ef1fb36c8116b5fe |
|
MD5 | ac958ae813fb2b4b963fdad0ec1a51cd |
|
BLAKE2b-256 | d756a1c99c033992d687d40f0b0d8fac34c58c33bf3f324dfd0ba2c9ec401266 |