General deep learing utility library
Project description
DeepZensols Deep Learning Framework
This deep learning library was designed to provide consistent and reproducible results.
- See the full documentation.
- Paper on arXiv.
Features:
- Easy to configure and framework to allow for programmatic debugging of neural networks.
- Reproducibility of results
- All random seed state is persisted in the trained model files.
- Persisting of keys and key order across train, validation and test sets.
- Analysis of results with complete metrics available.
- A vectorization framework that allows for pickling tensors.
- Additional layers:
- Full BiLSTM-CRF and stand-alone CRF implementation using easy to configure constituent layers.
- Easy to configure N [deep convolution layer] with automatic dimensionality calculation and configurable pooling and batch centering.
- Convolutional layer factory with dimensionality calculation.
- Recurrent layers that abstracts RNN, GRU and LSTM.
- N deep linear layers.
- Each layer's configurable with activation, dropout and batch normalization.
- Pandas integration to data load, easily manage vectorized features, and report results.
- Multi-process for time consuming CPU feature vectorization requiring little to no coding.
- Resource and tensor deallocation with memory management.
- Real-time performance and loss metrics with plotting while training.
- Thorough unit test coverage.
- Debugging layers using easy to configure Python logging module and control points.
Much of the code provides convenience functionality to PyTorch. However, there is functionality that could be used for other deep learning APIs.
Documentation
See the full documentation.
Obtaining
The easiest way to install the command line program is via the pip
installer:
pip3 install zensols.deeplearn
Binaries are also available on pypi.
Workflow
This package provides a workflow for processing features, training and then testing a model. A high level outline of this process follows:
- Container objects are used to represent and access data as features.
- Instances of data points wrap the container objects.
- Vectorize the features of each data point in to tensors.
- Store the vectorized tensor features to disk so they can be retrieved quickly and frequently.
- At train time, load the vectorized features in to memory and train.
- Test the model and store the results to disk.
To jump right in, see the examples section. However, it is better to peruse the in depth explanation with the Iris example code follows:
- The initial data processing, which includes data representation to batch creation.
- Creating and configuring the model.
- Using a facade to train, validate and test the model.
- Analysis of results, including training/validation loss graphs and performance metrics.
Examples
The Iris example (also see the Iris example configuration) is the most basic example of how to use this framework. This example is detailed in the workflow documentation in detail.
There are also examples in the form of Juypter notebooks as well, which include the:
- Iris notebook data set, which is a small data set of flower dimensions as a three label classification,
- MNIST notebook for the handwritten digit data set,
- debugging notebook.
Attribution
This project, or example code, uses:
- PyTorch as the underlying framework.
- Branched code from Torch CRF for the CRF class.
- pycuda for Python integration with CUDA.
- scipy for scientific utility.
- Pandas for prediction output.
- matplotlib for plotting loss curves.
Corpora used include:
Torch CRF
The CRF class was taken and modified from Kemal Kurniawan's pytorch_crf
GitHub repository. See the README.md
module documentation for more
information. This module was forked pytorch_crf with modifications.
However, the modifications were not merged and the project appears to be
inactive.
Important: This project will change to use it as a dependency pending
merging of the changes needed by this project. Until then, it will remain as a
separate class in this project, which is easier to maintain as the only
class/code is the CRF
class.
The pytorch_crf repository uses the same license as this repository, which the MIT License. For this reason, there are no software/package tainting issues.
See Also
The zensols deepnlp project is a deep learning utility library for natural language processing that aids in feature engineering and embedding layers that builds on this project.
Citation
If you use this framework in your research you must cite with:
@article{Landes_DiEugenio_Caragea_2021,
title={DeepZensols: Deep Natural Language Processing Framework},
url={http://arxiv.org/abs/2109.03383},
note={arXiv: 2109.03383},
journal={arXiv:2109.03383 [cs]},
author={Landes, Paul and Di Eugenio, Barbara and Caragea, Cornelia},
year={2021},
month={Sep}
}
Community
Please star the project and let me know how and where you use this API. Contributions as pull requests, feedback and any input is welcome.
Changelog
An extensive changelog is available here.
License
Copyright (c) 2020 - 2021 Paul Landes
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distributions
Built Distributions
Hashes for zensols.deeplearn-0.1.5-py3.9.egg
Algorithm | Hash digest | |
---|---|---|
SHA256 | e82366e5f7bf7cc4868c2c6483c616d6c72291676bbb82910e7dd0af16151328 |
|
MD5 | d725feed10dc8ee8feea346b10dff326 |
|
BLAKE2b-256 | 7d89c2479f05f25e4b09ef33fb260cc078f317c7c9318585b66cb7234e3d4efc |
Hashes for zensols.deeplearn-0.1.5-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 99230fa59e2d1847d341be58648047f75659d97d2a08b88e4863c4fb5a695d7c |
|
MD5 | 9ca647902beaad2a177abc5a054e3f3a |
|
BLAKE2b-256 | a2a41da21ea0bb19eb3e119ecbbbf3e7464f6c14621ca7a06370c814a074c673 |