Skip to main content

Preprocessed Paderborn Bearing Dataset for analysing multivariate motor current signals combined with a vibration signal

Project description

Paderborn Bearing Package

This package is created to extract and preprocess the Paderborn bearing datasets provided by the Paderborn University Faculty of Mechanical Engineering. This package is specifically designed to extract multivariate sensor readings from the MATLAB files containing different Bearing faults. The current version supports all operating systems, but keep in mind that you need different software packages installed depending on the operating system you use.

Dataset

The dataset consists of 25 damaged bearing states and 6 undamaged bearing states (Healthy condition). The damaged bearings are divided into artificial and real damages. Both of these damaged bearing states, two options are available where the damages are located, e.g., inner race (IR) and outer race (OR). One of the experiment types for the artificial OR condition is removed.

Installation

Installing the package can be done manually, but I recommend to install via pip since this is much easier by just copying the code below and paste it in your terminal/command prompt:

$ pip install paderborn-bearing

Please also check which dependencies are needed to use the package.

Use of the package

After the installation, the package can be imported in your specific session, for example in Jupyter Notebook;

from paderborn_bearing import *

Then the class in the package can be called as follows;

data = Paderborn("Artificial", 1024, "OR")

The called function results in three different objects:

  • data.motor_current, contains all the samples represented in a 3D-array for the two motor current sensor readings related to the specific experiment and fault name. The lay-out of the array is as follows; nr of time series, sequence length, amount of sensors.
  • data.vibrations, contains all the samples represented in a 3D-array for the vibration sensor related to the specific experiment and fault name. The lay-out of the array is as follows; nr of time series, sequence length, amount of sensors.
  • data.labels, labels representing different fault conditions.

Arguments

As can be seen in the previous code block, the package asks three arguments for extracting and preprocessing the Paderborn bearing datasets into usable objects. The following arguments need to be addressed when using Paderborn class:

  1. Experiment: A string value that indicates the specific name for the experiment to use. There are three options: "Artificial", "Healthy" and "Real".
  2. Length: an integer value that indicates the sequence length of every time series.
  3. Fault location: A string value that specifies the fault location of the bearing to narrow down the experiment's data. There are three options are available (Inner, Outer and Normal), "IR","OR" and "Normal". For the "Artificial" and "Real" options both "IR" and "OR" can be combined. For the "Heathly" option only the "Normal" condition can be used.

Dependencies

The following dependencies are needed to get the package installed. Because the MATLAB files need to be unpacked, we use different packages depending on the operating system.

  • numpy
  • patool (Windows)
  • scipy
  • subprocess (Mac OS/Linux)
  • tqdm
  • requests

MacOS/Linux

Since the package wants to unpack .rar files, you first need to install Homebrew in your terminal combined with Unar.

Paste this in a macOS Terminal to install Homebrew:

$/bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/Homebrew/install/HEAD/install.sh)"

Next, use the following command to install Unar:

$brew install unar

Additional information

The objects that are created are extracted from many MATLAB files, resulting in large-scale arrays. Therefore, I would like to emphasise to create every object separately, e.g., one experiment option + one fault location option to reduce the computation time of the data preprocessing. Keep in mind that the creation of one of these objects might take approximately 10 minutes since the arrays have to be stacked together across a wide variety of files. This results in large-scale data object of at least 3GB per object. I would therefore recommend to save the objects after preprocessing as numpy arrays, so that when you want to apply ML or DL algorithms, you only need to load in the objects (and potentially concatenate them).

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

paderborn_bearing-1.1.2.tar.gz (6.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

paderborn_bearing-1.1.2-py3-none-any.whl (6.4 kB view details)

Uploaded Python 3

File details

Details for the file paderborn_bearing-1.1.2.tar.gz.

File metadata

  • Download URL: paderborn_bearing-1.1.2.tar.gz
  • Upload date:
  • Size: 6.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.0

File hashes

Hashes for paderborn_bearing-1.1.2.tar.gz
Algorithm Hash digest
SHA256 8dca25ce692cda01025cb8e741a3e1ca71a1683c095684926a2509e7b0cc75da
MD5 6cf442b43dd0ebb4c57319c26d93b735
BLAKE2b-256 d19335e3acb77d9aedf832cff32cedec47869634af48296af72ce22ced1f7c63

See more details on using hashes here.

File details

Details for the file paderborn_bearing-1.1.2-py3-none-any.whl.

File metadata

File hashes

Hashes for paderborn_bearing-1.1.2-py3-none-any.whl
Algorithm Hash digest
SHA256 33346eae323a01b9726a680242a9368e07e34917f4c293631781aa99d0d48aee
MD5 04fc89bd95bbe94319ed8b338c1700b7
BLAKE2b-256 95710e8848cf43aff6e97644cbfa371730cf1c8267dd6e63475f6da8b0d7e528

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page