Skip to main content

Processes the data collected from the instrumented bicycle.

Project description

Description

This program is setup to process the raw data signals collected from the Davis Instrumented Bicycle’s data acquisition system (i.e. the output of BicycleDAQ). See [Moore2012] for details of the system and experiments.

License

BSD 2-Clause License, see LICENSE.txt.

Citation

If you make use of this data we kindly request that you cite our work, either [Moore2012], the software DOI, and/or other relevant references.

Dependencies

Installation

For ease of setup we recommend setting up a conda environment:

$ conda create -n bdp numpy scipy matplotlib "pytables<3.0" pyyaml
$ source activate bdp

The remaining dependencies need to be installed with pip:

(bdp)$ pip install "uncertainties>2.0.0" "DynamicistToolKit>=0.3.4"
(bdp)$ pip install "yeadon>=1.1.1" "BicycleParameters>=0.2.0"

And finally, this package:

(bdp)$ pip install BicycleDataProcessor

Usage

Load the prebuilt database file

The simplest way to get started with the data is to download the database file from:

$ wget http://files.figshare.com/1710608/instrumented_bicycle_raw_data_h5.tar.bz2
$ tar -jxvf instrumented_bicycle_raw_data_h5.tar.bz2

And also the bicycle parameter data:

$ wget http://files.figshare.com/1710525/bicycle_parameters.tar.gz
$ tar -zxvf bicycle_parameters.tar.gz
$ rm bicycle_parameters.tar.gz

In your working directory, create a bdp-defaults.cfg and change pathToDatabase and pathToParameters to point to the downloaded and unzipped database file and the bicycle-parameters data directory, respectively. See the example-bdp-defaults.cfg for reference. This file follows the standard Python configuration file format.

Interact with the data

Open a Python command prompt and import the module:

>>> import bicycledataprocessor as bdp

First load the database:

>>> dataset = bdp.DataSet()

Now load a run:

>>> run = bdp.Run('00105', dataset)

Check to make sure the data was properly time synchronized:

>>> run.verify_time_sync()

The graph that appears shows the mostly downward acceleration signals from the two accelerometers. These signals are used to synchronize the NI USB-2008 and the VN-100 data. If these do not match, then the synchronization algorithm didn’t not work and the data may be unusable.

The run has a lot of data associated with it. Firstly, you can print a subset of the meta data with:

>>> print(run)

The complete meta data is stored in a dictionary:

>>> run.metadata

The raw data for each sensor is stored in a dictionary and can be accessed by:

>>> run.rawSignals

The data for each sensor with calibration scaling can be accessed by:

>>> run.calibratedSignals

The data for each sensor after truncation based on the time synchronization can be accessed with:

>>> run.truncatedSignals

The data for each computed signal is also stored in a dictionary:

>>> run.computedSignals

The data for each task signal is also stored in a dictionary:

>>> run.taskSignals

The taskSignals can be plotted:

>>> run.taskSignals.keys() # see a list of options
>>> run.plot('SteerAngle', 'RollAngle', 'PullForce')

Export the computed signals as a mat file with:

>>> run.export('mat')

Build the HDF5 file from raw data

The second option would be to build the database with the raw data from BicycleDAQ. BicycleDAQ stores the raw data from trials and calibrations as Matlab mat files. Then use this module to create the database and fill it with the data.

The raw trial data can downloaded as so:

$ wget -O raw-trial-data.zip http://downloads.figshare.com/article/public/1164632
$ unzip -d raw-trial-data raw-trial-data.zip
$ rm raw-trial-data.zip

The raw calibration files:

$ wget -O raw-calibration-data.zip http://downloads.figshare.com/article/public/1164630
$ unzip -d raw-calibration-data raw-calibration-data.zip
$ rm raw-calibration-data.zip

And the additional corrupt trial file:

$ wget -O data-corruption.csv http://files.figshare.com/1696860/data_corruption.csv

Make sure your bdp-defaults.cfg paths point to the correct directories for the run mat files (pathToRunMat), calibration mat files (pathToCalibMat), the corrupt data file (data-corruption.csv). Optionally the paths can be set as arguments to DataSet().

Now create an empty database file in the current directory (or to the path specified in bdp-defaults.cfg if you’ve done that).:

$ python
>>> import bicycledataprocessor as bdp
>>> dataset = bdp.DataSet()
>>> dataset.create_database()

Now, fill the database with the data.:

>>> dataset.fill_all_tables()

The will take a little time to populate the database.

Warnings

  • The roll angle is not guaranteed to be calibrated in some of the early pavilion runs. Caution should be used.

  • The first set of pavilion runs with Luke and Charlie are mostly corrupt, beware. The corruption column in the runTable specifies which runs are corrupt.

  • The yaw angle and lateral deviation values depend on integrating the yaw rate. This seems to work for runs that have signals centered around zero, but may be wrong for others. (There are plans to fix this for all runs.)

Grant Information

This material is partially based upon work supported by the National Science Foundation under Grant No. 0928339. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation.

References

[Moore2012] (1,2)

Moore, J. K. Human Control of a Bicycle. University of California, Davis. 2012.

Release Notes

0.1.0

  • Initial PyPi release.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

BicycleDataProcessor-0.1.0.tar.gz (42.2 kB view details)

Uploaded Source

File details

Details for the file BicycleDataProcessor-0.1.0.tar.gz.

File metadata

File hashes

Hashes for BicycleDataProcessor-0.1.0.tar.gz
Algorithm Hash digest
SHA256 590f14c85dc2227d9e7457a265851359c94d5d178d19f8440de97def6599ee5c
MD5 6d286b65ed7f3310d2bdb6ebb040494c
BLAKE2b-256 f063e9815948a4d0fe19c20118e249529e212f070a5cbc9fe3700bdb0231b35b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page