A linear memory CUDA Time Warp Edit Distance algorithm.
Project description
cuTWED v2.0.0.
|Made With Python|
|DOI badge| |GPLv3 license| |GitHub release| |Open Source Love svg1|
A linear memory CUDA algorithm for solving Time Warp Edit Distance.
About
This is a novel parallel implementation of Time Warp Edit Distance.
The original algorithm was described by Marteau in:
Time Warp Edit Distance with Stiffness Adjustment for Time Series Matching (2009)
.
A variant of the code from that paper is included in
reference_implementation
, and was used to compare results. There is
also a wiki page <https://en.wikipedia.org/wiki/Time_Warp_Edit_Distance>
__.
The original TWED algorithm is O(n^2)
in time and space. This
algorithm is roughly O(n * n/p)
in time for p (CUDA) cores. Most
importantly, cuTWED is linear in memory, using storage for roughly
6*nA + 6*nB
elements.
In the process of understanding the dynamic program data dependencies in
order to parallelize towards the CUDA architecture, I devised a method
with improved memory access patterns and the ability to massively
parallelize over large problems. This is accomplished by a procession of
a three diagonal band moving across the dynamic program matrix in
nA+nB-1
steps. No O(n^2)
matrix is required anywhere. Note, this
is not an approximation method. The full and complete TWED dynamic
program computation occurs with linear storage.
The code is provided in a library that has methods for double
and
float
precision. It admits input time series in R^N
as arrays of
N-dimensional arrays in C-order (time is the slow moving axis).
For typical problems computable by the original TWED implementation, utilizing cuTWED and thousands of CUDA cores achieves great speedups. For common problems speedups are one to two orders of magnitude, capable of achieving 200x acceleration on a P100 GPUs in doubles. More so, the linear memory footprint allows for the computation of previously intractable problems. Large problems, large systems of inputs can be computed much more effectively now.
Some speed comparisons and a more formal paper will follow.
Reference Implementation
Marteau's original code with a some minor changes has been included in
this package. It is built both as a C library and part of the Python
package ``cuTWED.ctwed``. The minor changes are an extra argument
``dimension`` to admit ``R^N`` inputs, and more common handling of norm
(nth-root). These modications were made to facilitate refence testing
cuTWED and also make the original code more general.
``ctwed`` is also included for users without a CUDA card who find other
implementations too slow.
Getting Started
---------------
For many users the prepackaged (linux) Python distribution is the
simplest way to get the code. If you have CUDA10/11 and a Python 3.6+
manylinux compatible installation you can try the prepackged wheels with:
`pip install cuTWED`
For other situations, or users seeking maximum performance, instructions
follow for building the core CUDA libray, installing, and building
the Python bindings from source.
Requirements
~~~~~~~~~~~~
For the CUDA code you will need NVCC, a CUDA capable card and CMake.
Otherwise, the CUDA code has no dependencies.
If you do not have CMake or it is too old, a lot of people just pip
install it ``pip install cmake>=3.11``. Otherwise you'll need to
refer to their (Kitware) docs for your situation.
For the Python binding ``pip`` manages the specific depends and
installation of the Python interface after you have built the main CUDA
C library. Generally requires ``numpy``, ``pycuda``, and ``cffi``. I
recommend you use virtualenv or conda to manage Python.
Building
~~~~~~~~
Building has two stages.
First the CUDA C library is built and installed.
Second the Python bindings (if desired) are built on top of that.
The CUDA C library may be permanently installed to your system,
in a standard fashion, with some customization via CMake.
Alternatively a manual local install option is described below.
Building the core CUDA C library
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
Note you may customize the call to ``cmake`` below with flags like
``-DCMAKE_INSTALL_PREFIX=/opt/``, or other flags you might require.
::
# git a copy of the source code
git clone https://github.com/garrettwrong/cuTWED
cd cuTWED
# setup a build area
mkdir build
cd build
cmake ../ # configures/generates makefiles
# make
make -j # builds the software
This should create several files in the ``build`` directory including
``libcuTWED.so``, ``*.h`` headers, and some other stuff. To install to
your system permanently (may require sudo):
::
make install
If you would rather just want to temporarily have the library
available on your linux machine you can just use the LD path.
This makes no changes to your system outside the repo and this shell process.
Assuming you are still in your `build` directory, add the library to your path.
Users may decide to add a similar line permanently to their `.bashrc` or equivalent.
::
export LD_LIBRARY_PATH=$PWD:$LD_LIBRARY_PATH
Python
^^^^^^
Once you have the CUDA C library readied, we can use ``pip`` for the
Python bindings. From the root of the git repo:
::
pip install .
If you are planning to edit the code, you might prefer pip install in
local editable mode instead.
::
pip install -e .
Checking
~~~~~~~~
Assuming you have built both the CUDA library and the Python bindings,
you can run the unit test suite:
::
pytest
Development Testing
^^^^^^^^^^^^^^^^^^^
For developers there is a whole mess of configured tests which you can
run with:
::
tox --skip-missing-interpreters
I hope to improve this soon, but there are a *lot* of complication
running hybrid codes with free CI tools, and also packaging properly
with Python etc that need to be worked through. Some are most easily
addressed by using a managed CI host, but this is non free.... I suspect
this is largely why you do not see a plethera of free high performance
hybrid codes... perhaps a future project...
Using cuTWED in other programs
C/C++ ^^^^^
In C/C++ you should be able to include "cuTWED.h"
and link with the
shared library libcuTWED.so
. This is what I do in test.x
. The
public methods are extern C mangled and should be usable from both C and
C++ without issue.
Float (32bit) versions of all the public methods are included in the
shared library. They simply have an f
appended, for example,
twedf
is the float version of twed
. You may choose which one is
suitable for your application. I use floats in testf.x
.
There are currently two main ways to invoke the cuTWED alogorithm,
twed
and twed_dev
. First twed
is the most common way, where
you pass C arrays on the host, and the library manages device memory and
transfers for you.
Alternatively, if you are already managing GPU memory, you may use
twed_dev which expects pointers to memory that resides on the gpu. I
have also provided malloc, copy, and free helpers in case it makes sense
to reuse memory. See cuTWED.h
. All logic and size checks for such
advanced cases are expected to be owned by the user.
There is an additional batch method. Until I have a chance to write up
better documentation, you may find example use in test_batch
,
test_batch_dev
, and a small but respectable ML batch problem set in
test_synthetic_validation.py
.
I have included a Jupyter Notebook which demonstrates validation
using the UCI Pseudo Periodic Synthetic Time Series Data Set <http://archive.ics.uci.edu/ml/datasets/Pseudo+Periodic+Synthetic+Time+Series>
__. This is a much larger dataset.
Future plans include optimization and multi-gpu options for large batches..
Python ^^^^^^
::
from cuTWED import twed
For Python I have included basic pip installable Python bindings. I use
it in tests/test_basic.py
. If you are curious, these are implemented
by a cffi
backend which parses the C header. which is built for you
by setuptools
. The main Python interface is in cuTWED.py
. This
requires that you have built the library, and have it installed in a
location known to the system or available in your LD_LIBRARY_PATH
.
I have also wrapped up the GPU only memory methods in Python, using
PyCUDA gpuarrays. Examples in double and single precision are in
tests/test_basic_dev.py
.
::
from cuTWED import twed_dev
The batch interfaces are twed_batch
and twed_batch_dev
respectively. Currently it is doing a barbaric synchonization. I have a
branch using streams with events, but I need to validate it is robust
before I push it. That gives back about another 20% in batch mode
afaict.
If you want to run Marteau's C code from Python you can try ctwed
.
For (very) small problems you may find his original C code is faster.
Troubleshooting and Known Issues
This software is early in its life cycle. The following are known issues:
- Portability, I expect you have linux at this time.
- I have not had time to profile or optimize it, there are things I know to have improvements.
If you find an issue or bug with the code, please submit an issue. More details about this can be found in the contributing document.
License
GPLv3
Copyright 2020 Garrett Wright, Gestalt Group LLC
.. |DOI badge| image:: https://zenodo.org/badge/DOI/10.5281/zenodo.3842261.svg :target: https://doi.org/10.5281/zenodo.3842261 .. |GPLv3 license| image:: https://img.shields.io/badge/License-GPLv3-blue.svg :target: http://perso.crans.org/besson/LICENSE.html .. |GitHub release| image:: https://img.shields.io/github/release/garrettwrong/cuTWED.svg :target: https://GitHub.com/garrettwrong/cuTWED/releases/ .. |Open Source Love svg1| image:: https://badges.frapsoft.com/os/v1/open-source.svg?v=103 :target: https://github.com/ellerbrock/open-source-badges/ .. |Made With Python| image:: http://ForTheBadge.com/images/badges/made-with-python.svg :target: https://www.python.org/
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distributions
Hashes for cuTWED-2.0.0-cp39-cp39-manylinux2010_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | d18d8df5450921bbe54b058c019075d99d44c7ad504698ea048be3257312587a |
|
MD5 | ecbdfa671a54365a9707ddea20dc4cb1 |
|
BLAKE2b-256 | 0b23c8efa2da744981f8cfbd372783126e1061688773303befd16d01a4436633 |
Hashes for cuTWED-2.0.0-cp39-cp39-manylinux1_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 73bc6a532dcca95a233fd22b6ab3d0cd9ac0bb3b22fcbfcf0981b85cbfc82680 |
|
MD5 | a405007a531b6f60a5831bdda903cc24 |
|
BLAKE2b-256 | bd4e9fa8bddb875bbed1785ffa1ef26523ffd8f56816ab452b289d310c1fae9b |
Hashes for cuTWED-2.0.0-cp38-cp38-manylinux2010_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 59823a5f1be3739a33cea51b1d1c8bde4cd9af7c86942f2e7aa44ec178352179 |
|
MD5 | 5fb78b2d08a6862f2285e1fabc42d394 |
|
BLAKE2b-256 | 165d60be684e4ad88b81ea8a4bacace96f0da1efd0122a4b44c3acf7728a1df4 |
Hashes for cuTWED-2.0.0-cp38-cp38-manylinux1_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 7c9790683be70508ebd0d0a80ce845819b21178cb20a01020c475c55e058a410 |
|
MD5 | ce7514c01a7c24c88dc74e1bf75ff79c |
|
BLAKE2b-256 | 0ab3a6467ff7914a86214870932305806e49012b9923502921275b3c1e88c979 |
Hashes for cuTWED-2.0.0-cp37-cp37m-manylinux2010_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | ebfa90081604d243781453b1ed6b05c0c2cb7c5a17c2ab3737875ea036470959 |
|
MD5 | 2ab0823c1ecc152789cfc0a9a905a7eb |
|
BLAKE2b-256 | f448240242ee2284cd60d8181eb0b9ea63f3488c957009dd8570df884e9fe8a0 |
Hashes for cuTWED-2.0.0-cp37-cp37m-manylinux1_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 41da05c758082a69c3b23cba80067215f65b03279115a2309fd1da337da0d8e9 |
|
MD5 | cf000686b028a3a76a56bb24d9170a11 |
|
BLAKE2b-256 | f08dd31b51588ee28c811cb3055cb33b16f585ed18843c242cab98734d193f11 |
Hashes for cuTWED-2.0.0-cp36-cp36m-manylinux2010_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 3e45ec41a6f0a958cf74137202d55d40d30178f58ecc0ab5570942a8d07f2ff2 |
|
MD5 | b7a83f35ebfaad00ca15a26278fef550 |
|
BLAKE2b-256 | 0f06a8c54620764c7791aae82ce52dda27aeca506823d5de62d18d3387657041 |
Hashes for cuTWED-2.0.0-cp36-cp36m-manylinux1_x86_64.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | b3fadfdacf8fd9a986437cef66fedc63e233e273947d48d886830bdadbc252d8 |
|
MD5 | 95b6c42ac55294ffcc82202ff17b2700 |
|
BLAKE2b-256 | 6c091bd04abb0d03a9d6b947332255f5dfcce83ef4bf8c22c7671af23f75c7fc |