A collection of PyTorch implementations of neural network architectures and layers.
Project description
LabML Neural Networks
This is a collection of simple PyTorch implementation of various neural network architectures and layers. We will keep adding to this.
Transformers
Transformers module contains implementations for multi-headed attention and relative multi-headed attention.
Recurrent Highway Networks
This is the implementation for Recurrent Highway Networks.
LSTM
This is the implementation for LSTMs.
✅ Please create a Github issue if there’s something you’ld like to see implemented here.
Installation
pip install labml_nn
Links
Citing LabML
If you use LabML for academic research, please cite the library using the following BibTeX entry.
@misc{labml,
author = {Varuna Jayasiri, Nipun Wijerathne},
title = {LabML: A library to organize machine learning experiments},
year = {2020},
url = {https://lab-ml.com/},
}
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
labml_nn-0.4.2.tar.gz
(71.6 kB
view hashes)
Built Distribution
labml_nn-0.4.2-py3-none-any.whl
(109.3 kB
view hashes)