Skip to main content

Multimodal Sentiment Analysis Framework

Project description

MMSA

MMSA is a unified framework for Multimodal Sentiment Analysis.

Features

  • Train, test and compare multiple MSA models in a unified framework.
  • Supports 15 MSA models, including recent works.
  • Supports 3 MSA datasets: MOSI, MOSEI, and CH-SIMS.
  • Easy to use, provides Python APIs and commandline tools.
  • Experiment with fully customized multimodal features extracted by MMSA-FET toolkit.

1. Get Started

Note: From version 2.0, we packaged the project and uploaded it to PyPI in the hope of making it easier to use. If you don't like the new structure, you can always switch back to v_1.0 branch.

1.1 Use Python API

  • Run pip install MMSA in your python virtual environment.

  • Import and use in any python file:

    from MMSA import MMSA_run
    
    # run LMF on MOSI with default hyper parameters
    MMSA_run('lmf', 'mosi', seeds=[1111, 1112, 1113], gpu_ids=[0])
    
    # tune Self_mm on MOSEI with default hyper parameter range
    MMSA_run('self_mm', 'mosei', seeds=[1111], gpu_ids=[1])
    
    # run TFN on SIMS with altered config
    config = get_config_regression('tfn', 'mosi')
    config['post_fusion_dim'] = 32
    config['featurePath'] = '~/feature.pkl'
    MMSA_run('tfn', 'mosi', config=config, seeds=[1111])
    
    # run MTFN on SIMS with custom config file
    MMSA_run('mtfn', 'sims', config_file='./config.json')
    
  • For more detailed usage, please refer to APIs.

1.2 Use Commandline Tool

  • Run pip install MMSA in your python virtual environment.

  • Use from command line:

    # show usage
    $ python -m MMSA -h
    
    # train & test LMF on MOSI with default parameters
    $ python -m MMSA -d mosi -m lmf -s 1111 -s 1112
    
    # tune 50 times of TFN on MOSEI with custom config file & custom save dir
    $ python -m MMSA -d mosei -m tfn -t -tt 30 --model-save-dir ./models --res-save-dir ./results
    
    # train & test self_mm on SIMS with custom audio features & use gpu2
    $ python -m MMSA -d sims -m self_mm -Fa ./Features/Feature-A.pkl --gpu-ids 2
    
  • For more detailed usage, please refer to Commandline Arguments.

1.3 Clone & Edit the Code

  • Clone this repo and install requirements.
    $ git clone https://github.com/thuiar/MMSA
    
  • Edit the codes to your needs. See Code Structure for a basic review of our code structure.
  • After editing, run the following commands:
    $ cd MMSA-master # make sure you're in the top directory
    $ pip install .
    
  • Then run the code like above sections.
  • To further change the code, you need to re-install the package:
    $ pip uninstall MMSA
    $ pip install .
    
  • If you'd rather run the code without installation(like in v_1.0), please refer to Run Code without Installation.

2. Datasets

MMSA currently supports MOSI, MOSEI, and CH-SIMS dataset. Use the following links to download raw videos, feature files and label files. You don't need to download raw videos if you're not planning to run end-to-end tasks.

SHA-256 for feature files:

`MOSI/Processed/unaligned_50.pkl`:  `78e0f8b5ef8ff71558e7307848fc1fa929ecb078203f565ab22b9daab2e02524`
`MOSI/Processed/aligned_50.pkl`:    `d3994fd25681f9c7ad6e9c6596a6fe9b4beb85ff7d478ba978b124139002e5f9`
`MOSEI/Processed/unaligned_50.pkl`: `ad8b23d50557045e7d47959ce6c5b955d8d983f2979c7d9b7b9226f6dd6fec1f`
`MOSEI/Processed/aligned_50.pkl`:   `45eccfb748a87c80ecab9bfac29582e7b1466bf6605ff29d3b338a75120bf791`
`SIMS/Processed/unaligned_39.pkl`:  `c9e20c13ec0454d98bb9c1e520e490c75146bfa2dfeeea78d84de047dbdd442f`

MMSA uses feature files that are organized as follows:

{
    "train": {
        "raw_text": [],              # raw text
        "audio": [],                 # audio feature
        "vision": [],                # video feature
        "id": [],                    # [video_id$_$clip_id, ..., ...]
        "text": [],                  # bert feature
        "text_bert": [],             # word ids for bert
        "audio_lengths": [],         # audio feature lenth(over time) for every sample
        "vision_lengths": [],        # same as audio_lengths
        "annotations": [],           # strings
        "classification_labels": [], # Negative(0), Neutral(1), Positive(2). Deprecated in v_2.0
        "regression_labels": []      # Negative(<0), Neutral(0), Positive(>0)
    },
    "valid": {***},                  # same as "train"
    "test": {***},                   # same as "train"
}

Note: For MOSI and MOSEI, the pre-extracted text features are from BERT, different from the original glove features in the CMU-Multimodal-SDK.

Note: If you wish to extract customized multimodal features, please try out our MMSA-FET

3. Supported MSA Models

Type Model Name From Published
Single-Task TFN Tensor-Fusion-Network EMNLP 2017
Single-Task EF_LSTM MultimodalDNN ACL 2018 Workshop
Single-Task LF_DNN MultimodalDNN ACL 2018 Workshop
Single-Task LMF Low-rank-Multimodal-Fusion ACL 2018
Single-Task MFN Memory-Fusion-Network AAAI 2018
Single-Task Graph-MFN Graph-Memory-Fusion-Network ACL 2018
Single-Task MulT(without CTC) Multimodal-Transformer ACL 2019
Single-Task MFM MFM ICRL 2019
Multi-Task MLF_DNN MMSA ACL 2020
Multi-Task MTFN MMSA ACL 2020
Multi-Task MLMF MMSA ACL 2020
Single-Task BERT-MAG MAG-BERT ACL 2020
Single-Task MISA MISA ACMMM 2020
Single-Task SELF_MM Self-MM AAAI 2021
Single-Task MMIM MMIM EMNLP 2021
Single-Task BBFN (Work in Progress) BBFN ICMI 2021

4. Results

Baseline results are reported in results/result-stat.md

5. Citation

Please cite our paper if you find our work useful for your research:

@inproceedings{yu2020ch,
  title={CH-SIMS: A Chinese Multimodal Sentiment Analysis Dataset with Fine-grained Annotation of Modality},
  author={Yu, Wenmeng and Xu, Hua and Meng, Fanyang and Zhu, Yilin and Ma, Yixiao and Wu, Jiele and Zou, Jiyun and Yang, Kaicheng},
  booktitle={Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics},
  pages={3718--3727},
  year={2020}
}
@inproceedings{yu2021learning,
  title={Learning Modality-Specific Representations with Self-Supervised Multi-Task Learning for Multimodal Sentiment Analysis},
  author={Yu, Wenmeng and Xu, Hua and Yuan, Ziqi and Wu, Jiele},
  booktitle={Proceedings of the AAAI Conference on Artificial Intelligence},
  volume={35},
  number={12},
  pages={10790--10797},
  year={2021}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

MMSA-2.2.1.tar.gz (77.8 kB view details)

Uploaded Source

Built Distribution

MMSA-2.2.1-py3-none-any.whl (118.8 kB view details)

Uploaded Python 3

File details

Details for the file MMSA-2.2.1.tar.gz.

File metadata

  • Download URL: MMSA-2.2.1.tar.gz
  • Upload date:
  • Size: 77.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.15

File hashes

Hashes for MMSA-2.2.1.tar.gz
Algorithm Hash digest
SHA256 065b9ae1999816218e22ed54085c745ee95c93515aceaa6f8796f6ef70bfd6b9
MD5 6e240fe8031c5d83b8806f38e3a09c2f
BLAKE2b-256 bb95a122c7114d5f3c8d95ee22b5275154f2e40e01e232991d796c60fde8b6dd

See more details on using hashes here.

File details

Details for the file MMSA-2.2.1-py3-none-any.whl.

File metadata

  • Download URL: MMSA-2.2.1-py3-none-any.whl
  • Upload date:
  • Size: 118.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.15

File hashes

Hashes for MMSA-2.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 0dafd566f13c4a0064f6040f9f3788fca0dcc57813b81b51249732040a106728
MD5 71a55538b6c966a59118fe1e332b9126
BLAKE2b-256 8971219392da0ec89e5fcebcbdb22ddcead23b537817c49881a34e48771e93ad

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page