Skip to main content

General-purpose Multimodal Transformer with Linear Complexity Attention Mechanism.

Project description

LinMulT

License python pytorch

General-purpose Multimodal Transformer with Linear Complexity Attention Mechanism.

Setup

Install package from PyPI

pip install linmult

Install package for development

git clone https://github.com/fodorad/LinMulT
cd LinMulT
pip install -e .
pip install -U -r requirements.txt
python -m unittest

Similar projects using LinMulT

(2023) BlinkLinMulT

LinMulT is trained for blink presence detection and eye state recognition tasks. Our results demonstrate comparable or superior performance compared to state-of-the-art models on 2 tasks, using 7 public benchmark databases.

(2022) PersonalityLinMulT

LinMulT is trained for Big Five personality trait estimation using the First Impressions V2 dataset and sentiment estimation using the MOSI and MOSEI datasets.

Citation - BibTex

If you found our research helpful or influential please consider citing:

(2023) LinMulT for blink presence detection and eye state recognition:

@article{blinklinmult-fodor23,
  title = {BlinkLinMulT: Transformer-based Eye Blink Detection},
  author = {Fodor, {\'A}d{\'a}m and Fenech, Kristian and L{\H{o}}rincz, Andr{\'a}s},
  journal = {...}
  pages = {1--19},
  year = {2023}
}

(2022) LinMulT for personality trait and sentiment estimation:

@InProceedings{pmlr-v173-fodor22a,
  title = {Multimodal Sentiment and Personality Perception Under Speech: A Comparison of Transformer-based Architectures},
  author = {Fodor, {\'A}d{\'a}m and Saboundji, Rachid R. and Jacques Junior, Julio C. S. and Escalera, Sergio and Gallardo-Pujol, David and L{\H{o}}rincz, Andr{\'a}s},
  booktitle = {Understanding Social Behavior in Dyadic and Small Group Interactions},
  pages = {218--241},
  year = {2022},
  editor = {Palmero, Cristina and Jacques Junior, Julio C. S. and Clapés, Albert and Guyon, Isabelle and Tu, Wei-Wei and Moeslund, Thomas B. and Escalera, Sergio},
  volume = {173},
  series = {Proceedings of Machine Learning Research},
  month = {16 Oct},
  publisher = {PMLR},
  pdf = {https://proceedings.mlr.press/v173/fodor22a/fodor22a.pdf},
  url = {https://proceedings.mlr.press/v173/fodor22a.html}
}

Acknowledgement

The code is inspired by the following two materials:

Multimodal Transformer:

Linear Attention:

Contact

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

linmult-1.3.6.tar.gz (14.6 kB view details)

Uploaded Source

Built Distribution

linmult-1.3.6-py3-none-any.whl (17.1 kB view details)

Uploaded Python 3

File details

Details for the file linmult-1.3.6.tar.gz.

File metadata

  • Download URL: linmult-1.3.6.tar.gz
  • Upload date:
  • Size: 14.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: python-httpx/0.27.2

File hashes

Hashes for linmult-1.3.6.tar.gz
Algorithm Hash digest
SHA256 25e14fd49b9a15d95dcee8e9db9e210d9324dd5e138d14edabdcde6eae014bfd
MD5 5216a54865669b91484c9ace3022dba7
BLAKE2b-256 ffdc95936cc6a49c9bdebf58c62b8813a57d44e35d015ec5e2bf35f6d7f46a02

See more details on using hashes here.

File details

Details for the file linmult-1.3.6-py3-none-any.whl.

File metadata

  • Download URL: linmult-1.3.6-py3-none-any.whl
  • Upload date:
  • Size: 17.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: python-httpx/0.27.2

File hashes

Hashes for linmult-1.3.6-py3-none-any.whl
Algorithm Hash digest
SHA256 f038ef341b7adfbd955370a59931eb354e32b7677c89d26911dc89fb9a66e5fb
MD5 fefed970c0de1b540fe0ad29c1e9b26a
BLAKE2b-256 18e6e8e3476d655d38e10c940b8a5f256833e3eca24177b5f2480241d626dd81

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page