General-purpose Multimodal Transformer with Linear Complexity Attention Mechanism.
Project description
LinMulT
General-purpose Multimodal Transformer with Linear Complexity Attention Mechanism.
Setup
Install package from PyPI
pip install linmult
Install package for development
git clone https://github.com/fodorad/LinMulT
cd LinMulT
pip install -e .
pip install -U -r requirements.txt
python -m unittest
Similar projects using LinMulT
(2023) BlinkLinMulT
LinMulT is trained for blink presence detection and eye state recognition tasks. Our results demonstrate comparable or superior performance compared to state-of-the-art models on 2 tasks, using 7 public benchmark databases.
- paper: BlinkLinMulT: Transformer-based Eye Blink Detection (pdf, website)
- code: https://github.com/fodorad/BlinkLinMulT
(2022) PersonalityLinMulT
LinMulT is trained for Big Five personality trait estimation using the First Impressions V2 dataset and sentiment estimation using the MOSI and MOSEI datasets.
- paper: Multimodal Sentiment and Personality Perception Under Speech: A Comparison of Transformer-based Architectures (pdf, website)
- code: https://github.com/fodorad/PersonalityLinMulT
Citation - BibTex
If you found our research helpful or influential please consider citing:
(2023) LinMulT for blink presence detection and eye state recognition:
@article{blinklinmult-fodor23,
title = {BlinkLinMulT: Transformer-based Eye Blink Detection},
author = {Fodor, {\'A}d{\'a}m and Fenech, Kristian and L{\H{o}}rincz, Andr{\'a}s},
journal = {...}
pages = {1--19},
year = {2023}
}
(2022) LinMulT for personality trait and sentiment estimation:
@InProceedings{pmlr-v173-fodor22a,
title = {Multimodal Sentiment and Personality Perception Under Speech: A Comparison of Transformer-based Architectures},
author = {Fodor, {\'A}d{\'a}m and Saboundji, Rachid R. and Jacques Junior, Julio C. S. and Escalera, Sergio and Gallardo-Pujol, David and L{\H{o}}rincz, Andr{\'a}s},
booktitle = {Understanding Social Behavior in Dyadic and Small Group Interactions},
pages = {218--241},
year = {2022},
editor = {Palmero, Cristina and Jacques Junior, Julio C. S. and Clapés, Albert and Guyon, Isabelle and Tu, Wei-Wei and Moeslund, Thomas B. and Escalera, Sergio},
volume = {173},
series = {Proceedings of Machine Learning Research},
month = {16 Oct},
publisher = {PMLR},
pdf = {https://proceedings.mlr.press/v173/fodor22a/fodor22a.pdf},
url = {https://proceedings.mlr.press/v173/fodor22a.html}
}
Acknowledgement
The code is inspired by the following two materials:
Multimodal Transformer:
- paper: Multimodal Transformer for Unaligned Multimodal Language Sequences (1906.00295)
- code: https://github.com/yaohungt/Multimodal-Transformer
Linear Attention:
- paper: Transformers are RNNs: Fast Autoregressive Transformers with Linear Attention (2006.16236)
- code: https://github.com/idiap/fast-transformers
Contact
- Ádám Fodor (foauaai@inf.elte.hu) [website]
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file linmult-1.3.6.tar.gz
.
File metadata
- Download URL: linmult-1.3.6.tar.gz
- Upload date:
- Size: 14.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: python-httpx/0.27.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 25e14fd49b9a15d95dcee8e9db9e210d9324dd5e138d14edabdcde6eae014bfd |
|
MD5 | 5216a54865669b91484c9ace3022dba7 |
|
BLAKE2b-256 | ffdc95936cc6a49c9bdebf58c62b8813a57d44e35d015ec5e2bf35f6d7f46a02 |
File details
Details for the file linmult-1.3.6-py3-none-any.whl
.
File metadata
- Download URL: linmult-1.3.6-py3-none-any.whl
- Upload date:
- Size: 17.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: python-httpx/0.27.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f038ef341b7adfbd955370a59931eb354e32b7677c89d26911dc89fb9a66e5fb |
|
MD5 | fefed970c0de1b540fe0ad29c1e9b26a |
|
BLAKE2b-256 | 18e6e8e3476d655d38e10c940b8a5f256833e3eca24177b5f2480241d626dd81 |