Skip to main content

Sklearn flavored library containing numerous Activity of Daily Livings datasets, preprocessing methods, visualizations and models.

Project description

Activities of Daily Living - Machine Learning

Contains data preprocessing and visualization methods for ADL datasets.

PyPI version Download Stats Read the Docs (version) License

Activities of Daily living (ADLs) such as eating, working, sleeping and Smart Home device readings are recorded by inhabitants. Predicting resident activities based on the device event-stream allows for a range of applications, including automation, action recommendation and abnormal activity detection in the context of assisted living for elderly inhabitants. Pyadlml offers an easy way to fetch, visualize and preprocess common datasets.

!! Disclaimer !!

Package is still an alpha-version and under active development. As of now do not expect anything to work! APIs are going to change, stuff breaks and the documentation may lack behind. Nevertheless, feel free to take a look. The safest point to start is probably the API reference.

Last (stable) Release

$ pip install pyadlml

Latest Development Changes

$ git clone https://github.com/tcsvn/pyadlml
$ cd pyadlml
$ pip install .

Usage example

Simple

# Fetch dataset
from pyadlml.dataset import fetch_amsterdam
data = fetch_amsterdam()
df_devs, df_acts = data['devices'], data['activities']


# Plot the residents activity density over one day
from pyadlml.plot import plot_activity_density
fig = plot_activity_density(df_acts)
fig.show()


# Create a vector representing the state of all Smart Home devices
# at a certain time and discretize the events into 20 second bins
from pyadlml.preprocessing import Event2Vec, LabelMatcher
e2v = Event2Vec(encode='state', dt='20s')
states = e2v.fit_transform(df_devs)

# Label each datum with the corresponding activity.
# When an event matches no activity set the activity to "other"
lbls = LabelMatcher(other=True).fit_transform(df_acts, states)

# Extract numpy arrays without timestamps (1st column)
X, y = states.values[:,1:], lbls.values[:,1:]

# Proceed with machine learning stuff 
from sklearn.tree import DecisionTreeClassifier
clf = DecisionTreeClassifier()
clf.fit(X, y)
clf.score(X, y)
...

Less simple

from pyadlml.dataset import fetch_amsterdam
from pyadlml.constants import VALUE
from pyadlml.pipeline import Pipeline
from pyadlml.preprocessing import IndexEncoder, LabelMatcher, DropTimeIndex, \
                                  EventWindows, DropColumn
from pyadlml.model_selection import train_test_split
from pyadlml.model import DilatedModel
from pyadlml.dataset.torch import TorchDataset
from torch.utils.data import DataLoader
from torch.optim import Adam 
from torch.nn import functional as F

# Featch data and split into train/val/test based on time rather than #events
data = fetch_amsterdam()
X_train, X_val, X_test, y_train, y_val, y_test = train_test_split(
    data['devices'], data['activities'], \
    split=(0.6, 0.2, 0.2), temporal=True,
)

# Formulate all preprocessing steps using a sklearn styled pipeline 
pipe = Pipeline([
    ('enc', IndexEncoder()),            # Encode devices strings with indices
    ('drop_obs', DropColumn(VALUE)),    # Disregard device observations
    ('lbl', LabelMatcher(other=True)),  # Generate labels y  
    ('drop_time', DropTimeIndex()),     # Remove timestamps for x and y
    ('windows', EventWindows(           # Create sequences S with a sliding window
                  rep='many-to-one',    # Only one label y_i per sequence S_i
                  window_size=16,       # Each sequence contains 16 events 
                  stride=2)),           # A sequence is created every 2 events
    ('passthrough', 'passthrough')      # Do not use a classifier in the pipeline
])

# Create a dataset to sample from
dataset = TorchDataset(X_train, y_train, pipe) 
train_loader = DataLoader(dataset, batch_size=32, shuffle=True)
model = DilatedModel(
    n_features=14,       # Number of devices
    n_classes=8          # Number of activities
)
optimizer = Adam(model.parameters(), lr=3e-4)

# Minimal loop to overfit the data
for s in range(10000):
    for (Xb, yb) in train_loader:
        optimizer.zero_grad()
        logits = model(Xb)
        loss = F.cross_entropy(logits, yb)
        loss.backward()
        optimizer.step()

...

For more examples and how to use, please refer to the documentation.

Features

  • Access to 14 Datasets
  • Importing data from Home Assistant or Activity Assistant
  • Tools for data cleaning
    • Relabel activities and devices
    • Merge overlapping activities
    • Find and replace specific patterns in device signals
    • Interactive dashboard for data exploration
  • Various statistics and visualizations for devices, activities and their interaction
  • Preprocessing methods
    • Device encoder (index, state, changepoint, last_fired, ...)
    • Feature extraction (inter-event-times, intensity, time2vec, ...)
    • Sliding windows (event, temporal, explicit or (fuzzytime))
    • Many more ...
  • Cross validation iterators and pipeline adapted for ADLs
    • LeaveKDayOutSplit, TimeSeriesSplit
    • Conditional transformer: YTransformer, XorYTransformer, ...
  • Online metrics to compare models regardless of resample frequency
    • Accuracy, TPR, PPV, ConfMat, Calibration
  • Ready to use models (TODO)
    • RNNs
    • WaveNet
    • Transformer
  • Translate datasets to sktime formats

Supported Datasets

  • Amsterdam [1]
  • Aras [2]
  • Casas Aruba (2011) [3]
  • Casas Cairo [4]
  • Casas Milan (2009) [4]
  • Casas Tulum [4]
  • Casas Kyoto (2010) [4]
  • Kasteren House A,B,C [5]
  • MITLab [6]
  • UCI Adl Binary [8]
  • Chinokeeh [9]
  • Orange [TODO]

Examples, benchmarks and replications

The project includes (TODO) a ranked model leaderboard evaluated on the cleaned dataset versions. Additionaly, here is a useful list of awesome references (todo include link) to papers and repos related to ADLs and machine learning.

Contributing

  1. Fork it (https://github.com/tcsvn/pyadlml/fork)
  2. Create your feature branch (git checkout -b feature/fooBar)
  3. Commit your changes (git commit -am 'Add some fooBar')
  4. Push to the branch (git push origin feature/fooBar)
  5. Create a new Pull Request

Related projects

Support

Buy me a coffee

How to cite

If you are using pyadlml for publications consider citing the package

@software{pyadlml,
  author = {Christian Meier},
  title = {PyADLMl - Machine Learning library for Activities of Daily Living},    
  url = {https://github.com/tcsvn/pyadlml},
  version = {0.0.22-alpha},
  date = {2023-01-03}
}

Sources

Dataset

[1]: T.L.M. van Kasteren; A. K. Noulas; G. Englebienne and B.J.A. Kroese, Tenth International Conference on Ubiquitous Computing 2008
[2]: H. Alemdar, H. Ertan, O.D. Incel, C. Ersoy, ARAS Human Activity Datasets in Multiple Homes with Multiple Residents, Pervasive Health, Venice, May 2013.
[3,4]: WSU CASAS smart home project: D. Cook. Learning setting-generalized activity models for smart spaces. IEEE Intelligent Systems, 2011.
[5]: Transferring Knowledge of Activity Recognition across Sensor networks. Eighth International Conference on Pervasive Computing. Helsinki, Finland, 2010.
[6]: E. Munguia Tapia. Activity Recognition in the Home Setting Using Simple and Ubiquitous sensors. S.M Thesis
[7]: Activity Recognition in Smart Home Environments using Hidden Markov Models. Bachelor Thesis. Uni Tuebingen.
[8]: Ordonez, F.J.; de Toledo, P.; Sanchis, A. Activity Recognition Using Hybrid Generative/Discriminative Models on Home Environments Using Binary Sensors. Sensors 2013, 13, 5460-5477.
[9]: D. Cook and M. Schmitter-Edgecombe, Assessing the quality of activities in a smart environment. Methods of information in Medicine, 2009

Software

TODO add software used in TPPs

License

MIT © tcsvn

sdf

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pyadlml-0.0.8.0a0.tar.gz (272.9 kB view details)

Uploaded Source

Built Distribution

pyadlml-0.0.8.0a0-py3-none-any.whl (328.2 kB view details)

Uploaded Python 3

File details

Details for the file pyadlml-0.0.8.0a0.tar.gz.

File metadata

  • Download URL: pyadlml-0.0.8.0a0.tar.gz
  • Upload date:
  • Size: 272.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.8.10

File hashes

Hashes for pyadlml-0.0.8.0a0.tar.gz
Algorithm Hash digest
SHA256 7ea675c152511d13bdb29c0491171b77f15cbcbed66246f1c8bda3d2aab08c75
MD5 1d18c95d88aabf4be00f2224e22a9c33
BLAKE2b-256 8756956fc3d7fed6fc034350d71d8852b4b5538fb9c43dac4ef58de085c51172

See more details on using hashes here.

File details

Details for the file pyadlml-0.0.8.0a0-py3-none-any.whl.

File metadata

  • Download URL: pyadlml-0.0.8.0a0-py3-none-any.whl
  • Upload date:
  • Size: 328.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.2 CPython/3.8.10

File hashes

Hashes for pyadlml-0.0.8.0a0-py3-none-any.whl
Algorithm Hash digest
SHA256 9d75f04c83d89344849cddd1b5cb1c1f2ca75f7123be19b50df296ead8a27053
MD5 65dcd8f6eeec57f546f70b1d74cd1230
BLAKE2b-256 902b9676b1f8aa92be8146c7ad9a8b4c2aa735a84c95cb3f7aa84d42a88daf8b

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page