Skip to main content

Autoencoder based iterative modeling and multivariate time-series subsequence clustering algorithm (ABIMCA)

Project description

DOI Python 3.9

Autoencoder Based Iterative Modeling and Subsequence Clustering Algorithm (ABIMCA)

This repository contains the python code for the Autoencoder Based Iterative Modeling and Subsequence Clustering Algorithm (ABIMCA)[^koehn] which is a deep learning method to separate multivariate time-series data (MTSD) into subsequences. It is beneficial in a variety of fields, to cluster MTSD into smaller segments or subsequences in an unsupervised manner. The ability to filter measurement data based on specific subsequences can improve downstream development products such as anomaly detection or machine diagnosis in condition based maintenance (CbM) strategies. Our algorithm is specifically useful for MTSD generated by a mechatronic system in a transient environment. It can be used offline as well as online for streaming data. It utilizes recurrent neural network (RNN) based Autoencoders (AE) by iteratively training a Base Autoencoder (BAE), generating a segmentation score and saving the intermediate parameters of the BAE to recognize previously identified subsequences.

Autoencoder Based Iterative Modeling and Subsequence Clustering Algorithm (ABIMCA)

Usage

Package can be installed with pip

$ pip install abimca

or clone the repository, and cd into the directory. Then recommendation is to create a virtual environment after installing and using python 3.9 with pyenv

python -m venv .venv

activate the environment Linux: $ source .venv/bin/activate

Windows cmd: C:> .venv\Scripts\activate.bat

Then install with poetry

$ poetry install

Finally run the minimal example in main.py

$ python -m main

The above graphic example was generated with the run_lorenz.py script. The minimal example is as follows:

from abimca import SubsequenceIdentifier
import numpy as np


def main():
    # Generating random data. This will produce no class predictions or all points have the same class. For more reasonable results replace the data input with your mechatronic measurement data.
    # Number of datapoints (time-steps)
    n_p = 300
    # Number of dimensions or features
    dim = 5
    X = np.random.rand(n_p, dim)
    # Number of clusters
    n_c = 5
    y = np.random.randint(n_c, size=n_p)

    # Compute online clustering
    si = SubsequenceIdentifier(disable_progress_bar=False)
    si.fit(X)
    print(f"Label array from online fitting: \n{si.label_array}")

    # Compute offline clustering
    labels = si.predict(X)
    print(f"Label array from online fitting: \n{labels}")


if __name__ == "__main__":
    main()

References

[^koehn]: Köhne, J. et al. Autoencoder based iterative modeling and multivariate time-series subsequence clustering algorithm

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

abimca-0.2.5.tar.gz (17.6 kB view details)

Uploaded Source

Built Distribution

abimca-0.2.5-py3-none-any.whl (16.7 kB view details)

Uploaded Python 3

File details

Details for the file abimca-0.2.5.tar.gz.

File metadata

  • Download URL: abimca-0.2.5.tar.gz
  • Upload date:
  • Size: 17.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.13 CPython/3.9.13 Linux/5.15.0-46-generic

File hashes

Hashes for abimca-0.2.5.tar.gz
Algorithm Hash digest
SHA256 cb8d78700fd30eeea7ff6ff942a0039b39a4763adebb959046d69bff2d9e8249
MD5 7d0e413b3dc33775021051f9e4573665
BLAKE2b-256 0232ff7573bf00ef1236f0c253cbcdfe0d9d81d36c6688f631dd0f498d51264b

See more details on using hashes here.

File details

Details for the file abimca-0.2.5-py3-none-any.whl.

File metadata

  • Download URL: abimca-0.2.5-py3-none-any.whl
  • Upload date:
  • Size: 16.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.13 CPython/3.9.13 Linux/5.15.0-46-generic

File hashes

Hashes for abimca-0.2.5-py3-none-any.whl
Algorithm Hash digest
SHA256 91c59424dd650e49c09a42a8ca5a579c15b0a8f62f02c12b01e9fe7320ec7a0a
MD5 ee1331686788658bf7d89ca89c5d44c0
BLAKE2b-256 649bbecc9350c4526f454b1a3be8592272913cfb33d20eae57e1ba033cba969b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page