Skip to main content

A wake word detection toolkit

Project description

Howl

PyPI License: MPL 2.0

Wake word detection modeling for Firefox Voice, supporting open datasets like Google Speech Commands and Mozilla Common Voice.

Citation:

@inproceedings{tang-etal-2020-howl,
    title = "Howl: A Deployed, Open-Source Wake Word Detection System",
    author = "Tang, Raphael and Lee, Jaejun and Razi, Afsaneh and Cambre, Julia and Bicking, Ian and Kaye, Jofish and Lin, Jimmy",
    booktitle = "Proceedings of Second Workshop for NLP Open Source Software (NLP-OSS)",
    month = nov,
    year = "2020",
    publisher = "Association for Computational Linguistics",
    url = "https://www.aclweb.org/anthology/2020.nlposs-1.9",
    doi = "10.18653/v1/2020.nlposs-1.9",
    pages = "61--65"
}

Quickstart Guide

  1. Install PyAudio and PyTorch 1.5+ through your distribution's package system.

  2. Install Howl using pip

pip install howl
  1. To immediately use a pre-trained Howl model for inference, we provide the client API. The following example (also found under examples/hey_fire_fox.py) loads the "hey_fire_fox" pretrained model with a simple callback and starts the inference client.
from howl.client import HowlClient

def hello_callback(detected_words):
    print("Detected: {}".format(detected_words))

client = HowlClient()
client.from_pretrained("hey_fire_fox", force_reload=False)
client.add_listener(hello_callback)
client.start().join()

Training Guide

Installation

  1. git clone https://github.com/castorini/howl && cd howl

  2. Install PyTorch by following your platform-specific instructions.

  3. Install PyAudio and its dependencies through your distribution's package system.

  4. pip install -r requirements.txt -r requirements_training.txt (some apt packages might need to be installed)

Preparing a Dataset

In the example that follows, we describe how to train a custom detector for the word, "fire."

  1. Download a supported data source. We recommend Common Voice for its breadth and free license.
  2. To provide alignment for the data, install Montreal Forced Aligner (MFA) and download an English pronunciation dictionary.
  3. Create a positive dataset containing the keyword:
VOCAB='["fire"]' INFERENCE_SEQUENCE=[0] DATASET_PATH=data/fire-positive python -m howl_training.run.create_raw_dataset --negative-pct 0 -i ~/path/to/common-voice --positive-pct 100
  1. Create a negative dataset without the keyword:
VOCAB='["fire"]' INFERENCE_SEQUENCE=[0] DATASET_PATH=data/fire-negative python -m howl_training.run.create_raw_dataset --negative-pct 5 -i ~/path/to/common-voice --positive-pct 0
  1. Generate some mock alignment for the negative set, where we don't care about alignment:
DATASET_PATH=data/fire-negative python -m howl_training.run.attach_alignment --align-type stub
  1. Use MFA to generate alignment for the positive set:
mfa_align data/fire-positive/audio eng.dict pretrained_models/english.zip output-folder
  1. Attach the MFA alignment to the positive dataset:
DATASET_PATH=data/fire-positive python -m howl_training.run.attach_alignment --align-type mfa -i output-folder

Training and Running a Model

  1. Source the relevant environment variables for training the res8 model: source envs/res8.env.
  2. Train the model: python -m howl_training.run.train -i data/fire-negative data/fire-positive --model res8 --workspace workspaces/fire-res8.
  3. For the CLI demo, run python -m howl_training.run.demo --model res8 --workspace workspaces/fire-res8.

Pretrained Models

howl-models contains workspaces with pretrained models

To get the latest models, simply run git submodule update --init --recursive

VOCAB='[" hey","fire","fox"]' INFERENCE_SEQUENCE=[0,1,2] INFERENCE_THRESHOLD=0 NUM_MELS=40 MAX_WINDOW_SIZE_SECONDS=0.5 python -m howl_training.run.demo --model res8 --workspace howl-models/howl/hey-fire-fox

Reproducing Paper Results

First, follow the installation instructions in the quickstart guide.

Google Speech Commands

  1. Download the Google Speech Commands dataset and extract it.
  2. Source the appropriate environment variables: source envs/res8.env
  3. Set the dataset path to the root folder of the Speech Commands dataset: export DATASET_PATH=/path/to/dataset
  4. Train the res8 model: NUM_EPOCHS=20 MAX_WINDOW_SIZE_SECONDS=1 VOCAB='["yes","no","up","down","left","right","on","off","stop","go"]' BATCH_SIZE=64 LR_DECAY=0.8 LEARNING_RATE=0.01 python -m howl_training.run.pretrain_gsc --model res8

Hey Firefox

  1. Download the Hey Firefox corpus, licensed under CC0, and extract it.
  2. Download our noise dataset, built from Microsoft SNSD and MUSAN, and extract it.
  3. Source the appropriate environment variables: source envs/res8.env
  4. Set the noise dataset path to the root folder: export NOISE_DATASET_PATH=/path/to/snsd
  5. Set the firefox dataset path to the root folder: export DATASET_PATH=/path/to/hey_firefox
  6. Train the model: LR_DECAY=0.98 VOCAB='[" hey","fire","fox"]' USE_NOISE_DATASET=True BATCH_SIZE=16 INFERENCE_THRESHOLD=0 NUM_EPOCHS=300 NUM_MELS=40 INFERENCE_SEQUENCE=[0,1,2] MAX_WINDOW_SIZE_SECONDS=0.5 python -m howl_training.run.train --model res8 --workspace workspaces/hey-ff-res8

Hey Snips

  1. Download hey snips dataset
  2. Process the dataset to a format howl can load
VOCAB='["hey","snips"]' INFERENCE_SEQUENCE=[0,1] DATASET_PATH=data/hey-snips python -m howl_training.run.create_raw_dataset --dataset-type 'hey-snips' -i ~/path/to/hey_snips_dataset
  1. Generate some mock alignment for the dataset, where we don't care about alignment:
DATASET_PATH=data/hey-snips python -m howl_training.run.attach_alignment --align-type stub
  1. Use MFA to generate alignment for the dataset set:
mfa_align data/hey-snips/audio eng.dict pretrained_models/english.zip output-folder
  1. Attach the MFA alignment to the dataset:
DATASET_PATH=data/hey-snips python -m howl_training.run.attach_alignment --align-type mfa -i output-folder
  1. Source the appropriate environment variables: source envs/res8.env
  2. Set the noise dataset path to the root folder: export NOISE_DATASET_PATH=/path/to/snsd
  3. Set the noise dataset path to the root folder: export DATASET_PATH=/path/to/hey-snips
  4. Train the model: LR_DECAY=0.98 VOCAB='[" hey","snips"]' USE_NOISE_DATASET=True BATCH_SIZE=16 INFERENCE_THRESHOLD=0 NUM_EPOCHS=300 NUM_MELS=40 INFERENCE_SEQUENCE=[0,1] MAX_WINDOW_SIZE_SECONDS=0.5 python -m howl_training.run.train --model res8 --workspace workspaces/hey-snips-res8

Generating dataset for Mycroft-precise

howl also provides a script for transforming howl dataset to mycroft-precise dataset

VOCAB='[" hey","fire","fox"]' INFERENCE_SEQUENCE=[0,1,2] python -m howl_training.run.generate_precise_dataset --dataset-path /path/to/howl_dataset

Experiments

To verify the correctness of our implementation, we first train and evaluate our models on the Google Speech Commands dataset, for which there exists many known results. Next, we curate a wake word detection datasets and report our resulting model quality.

For both experiments, we generate reports in excel format. experiments folder includes sample outputs from the for each experiment and corresponding workspaces can be found here

commands_recognition

For command recognition, we train the four different models (res8, LSTM, LAS encoder, MobileNetv2) to detect twelve different keywords: “yes”, “no”, “up”, “down”, “left”, “right”, “on”, “off”, “stop”, “go”, unknown, or silence.

python -m howl_training.run.eval_commands_recognition --num_iterations n --dataset_path < path_to_gsc_datasets >

word_detection

In this experiment, we train our best commands recognition model, res8, for hey firefox and hey snips and evaluate them with different threashold.

Two different performance reports are generated, one with the clean audio and one with audios with noise

python -m howl_training.run.eval_wake_word_detection --num_models n --hop_size < number between 0 and 1 > --exp_type < hey_firefox | hey_snips > --dataset_path "x" --noiseset_path "y"

We also provide a script for generating ROC curve. exp_timestamp can be found from the reports generated from previous command

python -m howl_training.run.generate_roc --exp_timestamp < experiment timestamp > --exp_type < hey_firefox | hey_snips >

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

howl-0.1.2.tar.gz (34.6 kB view details)

Uploaded Source

Built Distribution

howl-0.1.2-py3-none-any.whl (45.2 kB view details)

Uploaded Python 3

File details

Details for the file howl-0.1.2.tar.gz.

File metadata

  • Download URL: howl-0.1.2.tar.gz
  • Upload date:
  • Size: 34.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.3.0 pkginfo/1.7.0 requests/2.25.1 setuptools/51.3.3 requests-toolbelt/0.9.1 tqdm/4.56.0 CPython/3.7.9

File hashes

Hashes for howl-0.1.2.tar.gz
Algorithm Hash digest
SHA256 9fe9685f2e00e521ea84e9b3a3b189b3510a717c0afb5fbb9020230f28d71f37
MD5 4c060e08b77dc6d06a6a80a5923ec4d6
BLAKE2b-256 4ece2fdb9eea3f3cf3ee7ceeb7484849426df2b74ddc3c97a73ee29ff66f1ce9

See more details on using hashes here.

File details

Details for the file howl-0.1.2-py3-none-any.whl.

File metadata

  • Download URL: howl-0.1.2-py3-none-any.whl
  • Upload date:
  • Size: 45.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.3.0 pkginfo/1.7.0 requests/2.25.1 setuptools/51.3.3 requests-toolbelt/0.9.1 tqdm/4.56.0 CPython/3.7.9

File hashes

Hashes for howl-0.1.2-py3-none-any.whl
Algorithm Hash digest
SHA256 c8b37a302accbab04bfd9a60560a862cc94ed2ca02c67c131ae6e2238e3a21c8
MD5 e37a45e886e930087749c236c42b45cc
BLAKE2b-256 eb1e89987301a46dfcc632a3a20302dad5c869f10ee1cc45b45e60bca6099777

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page