Skip to main content

Mother of All BCI Benchmarks

Project description

Mother of all BCI Benchmarks

banner

Build a comprehensive benchmark of popular BCI algorithms applied on an extensive list of freely available EEG datasets.

Disclaimer

This is an open science project that may evolve depending on the need of the community.

Build Status Code style: black

Welcome!

First and foremost, Welcome! :tada: Willkommen! :confetti_ball: Bienvenue! :balloon::balloon::balloon:

Thank you for visiting the Mother of all BCI Benchmark repository.

This document is a hub to give you some information about the project. Jump straight to one of the sections below, or just scroll down to find out more.

What are we doing?

The problem

  • Reproducible Research in BCI has a long way to go.
  • While many BCI datasets are made freely available, researchers do not publish code, and reproducing results required to benchmark new algorithms turns out to be more tricky than it should be.
  • Performances can be significantly impacted by parameters of the preprocessing steps, toolboxes used and implementation “tricks” that are almost never reported in the literature.

As a results, there is no comprehensive benchmark of BCI algorithm, and newcomers are spending a tremendous amount of time browsing literature to find out what algorithm works best and on which dataset.

The solution

The Mother of all BCI Benchmark allows to:

  • Build a comprehensive benchmark of popular BCI algorithms applied on an extensive list of freely available EEG datasets.
  • The code will be made available on github, serving as a reference point for the future algorithmic developments.
  • Algorithms can be ranked and promoted on a website, providing a clear picture of the different solutions available in the field.

This project will be successful when we read in an abstract “ … the proposed method obtained a score of 89% on the MOABB (Mother of All BCI Benchmarks), outperforming the state of the art by 5% ...”.

Who are we?

The founders of the Mother of all BCI Benchmarks are Alexander Barachant and Vinay Jayaram. This project is under the umbrella of NeuroTechX, the international community for NeuroTech enthusiasts. The project is currently maintained by Sylvain Chevallier.

What do we need?

You! In whatever way you can help.

We need expertise in programming, user experience, software sustainability, documentation and technical writing and project management.

We'd love your feedback along the way.

Our primary goal is to build a comprehensive benchmark of popular BCI algorithms applied on an extensive list of freely available EEG datasets, and we're excited to support the professional development of any and all of our contributors. If you're looking to learn to code, try out working collaboratively, or translate you skills to the digital domain, we're here to help.

Get involved

If you think you can help in any of the areas listed above (and we bet you can) or in any of the many areas that we haven't yet thought of (and here we're sure you can) then please check out our contributors' guidelines and our roadmap.

Please note that it's very important to us that we maintain a positive and supportive environment for everyone who wants to participate. When you join us we ask that you follow our code of conduct in all interactions both on and offline.

Contact us

If you want to report a problem or suggest an enhancement we'd love for you to open an issue at this github repository because then we can get right on it.

For a less formal discussion or exchanging ideas, you can also reach us on the Gitter channel or join our weekly office hours! This an open video meeting happening every Thursday at 18:30 GMT+1, please ask the link on the gitter channel. We are also on NeuroTechX slack #moabb channel.

Thank you

Thank you so much (Danke schön! Merci beaucoup!) for visiting the project and we do hope that you'll join us on this amazing journey to build a comprehensive benchmark of popular BCI algorithms applied on an extensive list of freely available EEG datasets.

Installation

A PyPi package will be available soon. For now, you need to fork or clone the repository and go to the downloaded directory, then run:

  1. install poetry (only once per machine):
    curl -sSL https://raw.githubusercontent.com/python-poetry/poetry/master/get-poetry.py | python -
    or checkout installation instruction or use conda forge version
  2. (Optional, skip if not sure) Disable automatical environment creation:
    poetry config virtualenvs.create false
  3. install all dependencies in one command (have to be run in project directory):
    poetry install

Requirements we use

see pyproject.toml file for full list of dependencies

Running

Verify Installation

To ensure it is running correctly, you can also run

python -m unittest moabb.tests

once it is installed.

Use MOABB

First, you could take a look at our tutorials, that cover the most important concepts and use cases. Also, we have a several examples available.

You might be interested in MOABB documentation

Supported datasets

The list of supported dataset can be found here : http://moabb.neurotechx.com/docs/datasets.html

Submit a new dataset

you can submit new dataset by mentioning it to this issue. The datasets currently on our radar can be seen [here] (https://github.com/NeuroTechX/moabb/wiki/Datasets-Support)

Architecture and main concepts

banner

there are 4 main concepts in the MOABB: the datasets, the paradigm, the evaluation, and the pipelines. In addition, we offer statistical and visualization utilities to simplify the workflow.

Datasets

A dataset handle and abstract low level access to the data. the dataset will takes data stored locally, in the format in which they have been downloaded, and will convert them into a MNE raw object. There are options to pool all the different recording sessions per subject or to evaluate them separately.

Paradigm

A paradigm defines how the raw data will be converted to trials ready to be processed by a decoding algorithm. This is a function of the paradigm used, i.e. in motor imagery one can have two-class, multi-class, or continuous paradigms; similarly, different preprocessing is necessary for ERP vs ERD paradigms.

Evaluations

An evaluation defines how we go from trials per subject and session to a generalization statistic (AUC score, f-score, accuracy, etc) -- it can be either within-recording-session accuracy, across-session within-subject accuracy, across-subject accuracy, or other transfer learning settings.

Pipelines

Pipeline defines all steps required by an algorithm to obtain predictions. Pipelines are typically a chain of sklearn compatible transformers and end with an sklearn compatible estimator. See Pipelines for more info.

Statistics and visualization

Once an evaluation has been run, the raw results are returned as a DataFrame. This can be further processed via the following commands to generate some basic visualization and statistical comparisons:

from moabb.analysis import analyze

results = evaluation.process(pipeline_dict)
analyze(results)

Citing MOABB and related publications

To cite MOABB, you could use the following paper:

Vinay Jayaram and Alexandre Barachant. "MOABB: trustworthy algorithm benchmarking for BCIs." Journal of neural engineering 15.6 (2018): 066011. DOI

If you publish a paper using MOABB, please contact us on gitter or open an issue, and we will add you paper to the dedicated wiki page.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

moabb-0.3.0.tar.gz (87.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

moabb-0.3.0-py3-none-any.whl (109.7 kB view details)

Uploaded Python 3

File details

Details for the file moabb-0.3.0.tar.gz.

File metadata

  • Download URL: moabb-0.3.0.tar.gz
  • Upload date:
  • Size: 87.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.5 CPython/3.6.13 Darwin/19.6.0

File hashes

Hashes for moabb-0.3.0.tar.gz
Algorithm Hash digest
SHA256 6e4cf42be96ed98511c721bdcd5a725ce16e5dddfa33cda8331e09ceba315288
MD5 20a063a5a438b2a1c87e23278d45ace5
BLAKE2b-256 93e684a5f69ab6569f2c18f3fcab7b42f5e3574c5c7ac428f71de381094e3edd

See more details on using hashes here.

File details

Details for the file moabb-0.3.0-py3-none-any.whl.

File metadata

  • Download URL: moabb-0.3.0-py3-none-any.whl
  • Upload date:
  • Size: 109.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.1.5 CPython/3.6.13 Darwin/19.6.0

File hashes

Hashes for moabb-0.3.0-py3-none-any.whl
Algorithm Hash digest
SHA256 224b9a24aafe9a6d8d3474f9c40b8d47c44d13f7fe3fbb16e88ea9c6ee3cd68b
MD5 af5bdc58aec518bb4be34d8c43ee5cbd
BLAKE2b-256 cf02c5060df989268923dee901c30a1b3e5e45ebc2941bc52ce7f8dea1b647c1

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page