Skip to main content

A library for running inference on a DeepSpeech model

Project description

Project DeepSpeech

Task Status

DeepSpeech is an open source Speech-To-Text engine, using a model trained by machine learning techniques based on Baidu's Deep Speech research paper. Project DeepSpeech uses Google's TensorFlow to make the implementation easier.

Usage

Pre-built binaries for performing inference with a trained model can be installed with pip3. Proper setup using a virtual environment is recommended, and you can find that documentation below.

A pre-trained English model is available for use and can be downloaded using the instructions below. Currently, only 16-bit, 16 kHz, mono-channel WAVE audio files are supported in the Python client.

Once everything is installed, you can then use the deepspeech binary to do speech-to-text on short (approximately 5-second long) audio files as such:

pip3 install deepspeech
deepspeech --model models/output_graph.pbmm --alphabet models/alphabet.txt --lm models/lm.binary --trie models/trie --audio my_audio_file.wav

Alternatively, quicker inference can be performed using a supported NVIDIA GPU on Linux. See the release notes to find which GPUs are supported. To run deepspeech on a GPU, install the GPU specific package:

pip3 install deepspeech-gpu
deepspeech --model models/output_graph.pbmm --alphabet models/alphabet.txt --lm models/lm.binary --trie models/trie --audio my_audio_file.wav

Please ensure you have the required CUDA dependency.

See the output of deepspeech -h for more information on the use of deepspeech. (If you experience problems running deepspeech, please check required runtime dependencies).

Table of Contents

Prerequisites

Getting the code

Install Git Large File Storage either manually or through a package-manager if available on your system. Then clone the DeepSpeech repository normally:

git clone https://github.com/mozilla/DeepSpeech

Getting the pre-trained model

If you want to use the pre-trained English model for performing speech-to-text, you can download it (along with other important inference material) from the DeepSpeech releases page. Alternatively, you can run the following command to download and unzip the model files in your current directory:

wget https://github.com/mozilla/DeepSpeech/releases/download/v0.4.1/deepspeech-0.4.1-models.tar.gz
tar xvfz deepspeech-0.4.1-models.tar.gz

Using the model

There are three ways to use DeepSpeech inference:

CUDA dependency

The GPU capable builds (Python, NodeJS, C++ etc) depend on the same CUDA runtime as upstream TensorFlow. Currently with TensorFlow r1.12 it depends on CUDA 9.0 and CuDNN v7.2.

Using the Python package

Pre-built binaries which can be used for performing inference with a trained model can be installed with pip3. You can then use the deepspeech binary to do speech-to-text on an audio file:

For the Python bindings, it is highly recommended that you perform the installation within a Python 3.5 or later virtual environment. You can find more information about those in this documentation.

We will continue under the assumption that you already have your system properly setup to create new virtual environments.

Create a DeepSpeech virtual environment

In creating a virtual environment you will create a directory containing a python3 binary and everything needed to run deepspeech. You can use whatever directory you want. For the purpose of the documentation, we will rely on $HOME/tmp/deepspeech-venv. You can create it using this command:

$ virtualenv -p python3 $HOME/tmp/deepspeech-venv/

Once this command completes successfully, the environment will be ready to be activated.

Activating the environment

Each time you need to work with DeepSpeech, you have to activate this virtual environment. This is done with this simple command:

$ source $HOME/tmp/deepspeech-venv/bin/activate

Installing DeepSpeech Python bindings

Once your environment has been set-up and loaded, you can use pip3 to manage packages locally. On a fresh setup of the virtualenv, you will have to install the DeepSpeech wheel. You can check if deepspeech is already installed with pip3 list.

To perform the installation, just use pip3 as such:

$ pip3 install deepspeech

If deepspeech is already installed, you can update it as such:

$ pip3 install --upgrade deepspeech

Alternatively, if you have a supported NVIDIA GPU on Linux, you can install the GPU specific package as follows:

$ pip3 install deepspeech-gpu

See the release notes to find which GPUs are supported. Please ensure you have the required CUDA dependency.

You can update deepspeech-gpu as follows:

$ pip3 install --upgrade deepspeech-gpu

In both cases, pip3 should take care of installing all the required dependencies. After installation has finished, you should be able to call deepspeech from the command-line.

Note: the following command assumes you downloaded the pre-trained model.

deepspeech --model models/output_graph.pbmm --alphabet models/alphabet.txt --lm models/lm.binary --trie models/trie --audio my_audio_file.wav

The arguments --lm and --trie are optional, and represent a language model.

See client.py for an example of how to use the package programatically.

Using the command-line client

To download the pre-built binaries for the deepspeech command-line client, use util/taskcluster.py:

python3 util/taskcluster.py --target .

or if you're on macOS:

python3 util/taskcluster.py --arch osx --target .

also, if you need some binaries different than current master, like v0.2.0-alpha.6, you can use --branch:

python3 util/taskcluster.py --branch "v0.2.0-alpha.6" --target "."

The script taskcluster.py will download native_client.tar.xz (which includes the deepspeech binary and associated libraries) and extract it into the current folder. Also, taskcluster.py will download binaries for Linux/x86_64 by default, but you can override that behavior with the --arch parameter. See the help info with python util/taskcluster.py -h for more details. Specific branches of DeepSpeech or TensorFlow can be specified as well.

Note: the following command assumes you downloaded the pre-trained model.

./deepspeech --model models/output_graph.pbmm --alphabet models/alphabet.txt --lm models/lm.binary --trie models/trie --audio audio_input.wav

See the help output with ./deepspeech -h and the native client README for more details.

Using the Node.JS package

You can download the Node.JS bindings using npm:

npm install deepspeech

Alternatively, if you're using Linux and have a supported NVIDIA GPU, you can install the GPU specific package as follows:

npm install deepspeech-gpu

See the release notes to find which GPUs are supported. Please ensure you have the required CUDA dependency.

See client.js for an example of how to use the bindings. Or download the wav example.

Installing bindings from source

If pre-built binaries aren't available for your system, you'll need to install them from scratch. Follow these native_client installation instructions.

Third party bindings

In addition to the bindings above, third party developers have started to provide bindings to other languages:

Training

Installing prerequisites for training

Install the required dependencies using pip3:

cd DeepSpeech
pip3 install -r requirements.txt

You'll also need to install the ds_ctcdecoder Python package. ds_ctcdecoder is required for decoding the outputs of the deepspeech acoustic model into text. You can use util/taskcluster.py with the --decoder flag to get a URL to a binary of the decoder package appropriate for your platform and Python version:

pip3 install $(python3 util/taskcluster.py --decoder)

This command will download and install the ds_ctcdecoder package. If you prefer building the binaries from source, see the native_client README file. You can override the platform with --arch if you want the package for ARM7 (--arch arm) or ARM64 (--arch arm64).

Recommendations

If you have a capable (NVIDIA, at least 8GB of VRAM) GPU, it is highly recommended to install TensorFlow with GPU support. Training will be significantly faster than using the CPU. To enable GPU support, you can do:

pip3 uninstall tensorflow
pip3 install 'tensorflow-gpu==1.13.1'

Please ensure you have the required CUDA dependency.

Common Voice training data

The Common Voice corpus consists of voice samples that were donated through Mozilla's Common Voice Initiative.

We provide an importer (bin/import_cv.py) which automates downloading and preparing the Common Voice corpus as such:

bin/import_cv.py path/to/target/directory

If you already downloaded Common Voice from here, simply run bin/import_cv.py on the directory where the corpus is located. The importer will detect that you've already downloaded the data and immediately proceed to unpackaging and importing. If you haven't downloaded the data already, bin/import_cv.py will download it for you and save to the path you've specified.

Please be aware that training with the Common Voice corpus archive requires at least 70GB of free disk space and quite some time to conclude. As this process creates a huge number of small files, using an SSD drive is highly recommended. If the import script gets interrupted, it will try to continue from where it stopped the next time you run it. Unfortunately, there are some cases where it will need to start over. Once the import is done, the directory will contain a bunch of CSV files.

The following files are official user-validated sets for training, validating and testing:

  • cv-valid-train.csv
  • cv-valid-dev.csv
  • cv-valid-test.csv

The following files are the non-validated unofficial sets for training, validating and testing:

  • cv-other-train.csv
  • cv-other-dev.csv
  • cv-other-test.csv

cv-invalid.csv contains all samples that users flagged as invalid.

A sub-directory called cv_corpus_{version} contains the mp3 and wav files that were extracted from an archive named cv_corpus_{version}.tar.gz. All entries in the CSV files refer to their samples by absolute paths. So moving this sub-directory would require another import or tweaking the CSV files accordingly.

To use Common Voice data during training, validation and testing, you pass (comma separated combinations of) their filenames into --train_files, --dev_files, --test_files parameters of DeepSpeech.py.

If, for example, Common Voice was imported into ../data/CV, DeepSpeech.py could be called like this:

./DeepSpeech.py --train_files ../data/CV/cv-valid-train.csv --dev_files ../data/CV/cv-valid-dev.csv --test_files ../data/CV/cv-valid-test.csv

If you are brave enough, you can also include the other dataset, which contains not-yet-validated content:

./DeepSpeech.py --train_files ../data/CV/cv-valid-train.csv,../data/CV/cv-other-train.csv --dev_files ../data/CV/cv-valid-dev.csv --test_files ../data/CV/cv-valid-test.csv

Training a model

The central (Python) script is DeepSpeech.py in the project's root directory. For its list of command line options, you can call:

./DeepSpeech.py --helpfull

To get the output of this in a slightly better-formatted way, you can also look up the option definitions top DeepSpeech.py.

For executing pre-configured training scenarios, there is a collection of convenience scripts in the bin folder. Most of them are named after the corpora they are configured for. Keep in mind that the other speech corpora are very large, on the order of tens of gigabytes, and some aren't free. Downloading and preprocessing them can take a very long time, and training on them without a fast GPU (GTX 10 series recommended) takes even longer.

If you experience GPU OOM errors while training, try reducing the batch size with the --train_batch_size, --dev_batch_size and --test_batch_size parameters.

As a simple first example you can open a terminal, change to the directory of the DeepSpeech checkout and run:

./bin/run-ldc93s1.sh

This script will train on a small sample dataset called LDC93S1, which can be overfitted on a GPU in a few minutes for demonstration purposes. From here, you can alter any variables with regards to what dataset is used, how many training iterations are run and the default values of the network parameters.

Feel also free to pass additional (or overriding) DeepSpeech.py parameters to these scripts. Then, just run the script to train the modified network.

Each dataset has a corresponding importer script in bin/ that can be used to download (if it's freely available) and preprocess the dataset. See bin/import_librivox.py for an example of how to import and preprocess a large dataset for training with DeepSpeech.

If you've run the old importers (in util/importers/), they could have removed source files that are needed for the new importers to run. In that case, simply remove the extracted folders and let the importer extract and process the dataset from scratch, and things should work.

Checkpointing

During training of a model so-called checkpoints will get stored on disk. This takes place at a configurable time interval. The purpose of checkpoints is to allow interruption (also in the case of some unexpected failure) and later continuation of training without losing hours of training time. Resuming from checkpoints happens automatically by just (re)starting training with the same --checkpoint_dir of the former run.

Be aware however that checkpoints are only valid for the same model geometry they had been generated from. In other words: If there are error messages of certain Tensors having incompatible dimensions, this is most likely due to an incompatible model change. One usual way out would be to wipe all checkpoint files in the checkpoint directory or changing it before starting the training.

Exporting a model for inference

If the --export_dir parameter is provided, a model will have been exported to this directory during training. Refer to the corresponding README.md for information on building and running a client that can use the exported model.

Exporting a model for TFLite

If you want to experiment with the TF Lite engine, you need to export a model that is compatible with it, then use the --export_tflite flag. If you already have a trained model, you can re-export it for TFLite by running DeepSpeech.py again and specifying the same checkpoint_dir that you used for training, as well as passing --notrain --notest --export_tflite --export_dir /model/export/destination.

Making a mmap-able model for inference

The output_graph.pb model file generated in the above step will be loaded in memory to be dealt with when running inference. This will result in extra loading time and memory consumption. One way to avoid this is to directly read data from the disk.

TensorFlow has tooling to achieve this: it requires building the target //tensorflow/contrib/util:convert_graphdef_memmapped_format (binaries are produced by our TaskCluster for some systems including Linux/amd64 and macOS/amd64), use util/taskcluster.py tool to download, specifying tensorflow as a source and convert_graphdef_memmapped_format as artifact.

Producing a mmap-able model is as simple as:

$ convert_graphdef_memmapped_format --in_graph=output_graph.pb --out_graph=output_graph.pbmm

Upon sucessfull run, it should report about conversion of a non-zero number of nodes. If it reports converting 0 nodes, something is wrong: make sure your model is a frozen one, and that you have not applied any incompatible changes (this includes quantize_weights).

Distributed training across more than one machine

DeepSpeech has built-in support for distributed TensorFlow. To get an idea on how this works, you can use the script bin/run-cluster.sh for running a cluster with workers just on the local machine.

$ bin/run-cluster.sh --help
Usage: run-cluster.sh [--help] [--script script] [p:w:g] <arg>*

--help      print this help message
--script    run the provided script instead of DeepSpeech.py
p           number of local parameter servers
w           number of local workers
g           number of local GPUs per worker
<arg>*      remaining parameters will be forwarded to DeepSpeech.py or a provided script

Example usage - The following example will create a local DeepSpeech.py cluster
with 1 parameter server, and 2 workers with 1 GPU each:
$ run-cluster.sh 1:2:1 --epoch 10

Be aware that for the help example to be able to run, you need at least two CUDA capable GPUs (2 workers x 1 GPU). The script utilizes environment variable CUDA_VISIBLE_DEVICES for DeepSpeech.py to see only the provided number of GPUs per worker.

The script is meant to be a template for your own distributed computing instrumentation. Just modify the startup code for the different servers (workers and parameter servers) accordingly. You could use SSH or something similar for running them on your remote hosts.

Continuing training from a release model

If you'd like to use one of the pre-trained models released by Mozilla to bootstrap your training process (transfer learning, fine tuning), you can do so by using the --checkpoint_dir flag in DeepSpeech.py. Specify the path where you downloaded the checkpoint from the release, and training will resume from the pre-trained model.

For example, if you want to fine tune the entire graph using your own data in my-train.csv, my-dev.csv and my-test.csv, for three epochs, you can something like the following, tuning the hyperparameters as needed:

mkdir fine_tuning_checkpoints
python3 DeepSpeech.py --n_hidden 2048 --checkpoint_dir path/to/checkpoint/folder --epoch -3 --train_files my-train.csv --dev_files my-dev.csv --test_files my_dev.csv --learning_rate 0.0001

Note: the released models were trained with --n_hidden 2048, so you need to use that same value when initializing from the release models. Note as well the use of a negative epoch count -3 (meaning 3 more epochs) since the checkpoint you're loading from was already trained for several epochs.

Contact/Getting Help

There are several ways to contact us or to get help:

  1. FAQ - We have a list of common questions, and their answers, in our FAQ. When just getting started, it's best to first check the FAQ to see if your question is addressed.

  2. Discourse Forums - If your question is not addressed in the FAQ, the Discourse Forums is the next place to look. They contain conversations on General Topics, Using Deep Speech, and Deep Speech Development.

  3. IRC - If your question is not addressed by either the FAQ or Discourse Forums, you can contact us on the #machinelearning channel on Mozilla IRC; people there can try to answer/help

  4. Issues - Finally, if all else fails, you can open an issue in our repo.

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

If you're not sure about the file name format, learn more about wheel file names.

deepspeech-0.5.0a3-cp37-cp37m-manylinux1_x86_64.whl (15.2 MB view details)

Uploaded CPython 3.7m

deepspeech-0.5.0a3-cp37-cp37m-macosx_10_10_x86_64.whl (22.0 MB view details)

Uploaded CPython 3.7mmacOS 10.10+ x86-64

deepspeech-0.5.0a3-cp36-cp36m-manylinux1_x86_64.whl (15.2 MB view details)

Uploaded CPython 3.6m

deepspeech-0.5.0a3-cp36-cp36m-macosx_10_10_x86_64.whl (22.0 MB view details)

Uploaded CPython 3.6mmacOS 10.10+ x86-64

deepspeech-0.5.0a3-cp35-cp35m-manylinux1_x86_64.whl (15.2 MB view details)

Uploaded CPython 3.5m

deepspeech-0.5.0a3-cp35-cp35m-macosx_10_10_x86_64.whl (22.0 MB view details)

Uploaded CPython 3.5mmacOS 10.10+ x86-64

deepspeech-0.5.0a3-cp35-cp35m-linux_armv7l.whl (18.9 MB view details)

Uploaded CPython 3.5m

deepspeech-0.5.0a3-cp34-cp34m-manylinux1_x86_64.whl (15.2 MB view details)

Uploaded CPython 3.4m

deepspeech-0.5.0a3-cp34-cp34m-macosx_10_10_x86_64.whl (22.0 MB view details)

Uploaded CPython 3.4mmacOS 10.10+ x86-64

deepspeech-0.5.0a3-cp34-cp34m-linux_armv7l.whl (18.9 MB view details)

Uploaded CPython 3.4m

deepspeech-0.5.0a3-cp27-cp27mu-manylinux1_x86_64.whl (15.2 MB view details)

Uploaded CPython 2.7mu

deepspeech-0.5.0a3-cp27-cp27mu-macosx_10_10_x86_64.whl (22.0 MB view details)

Uploaded CPython 2.7mumacOS 10.10+ x86-64

deepspeech-0.5.0a3-cp27-cp27m-manylinux1_x86_64.whl (15.2 MB view details)

Uploaded CPython 2.7m

deepspeech-0.5.0a3-cp27-cp27m-macosx_10_10_x86_64.whl (22.0 MB view details)

Uploaded CPython 2.7mmacOS 10.10+ x86-64

File details

Details for the file deepspeech-0.5.0a3-cp37-cp37m-manylinux1_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp37-cp37m-manylinux1_x86_64.whl
  • Upload date:
  • Size: 15.2 MB
  • Tags: CPython 3.7m
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp37-cp37m-manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 6b89d5c1775a78ff422f72e081b3ffff4b07707b19de7edda2c6e703749ab34d
MD5 34b8271a4445590561a9c499862514bd
BLAKE2b-256 75dedc8bb5744a0f6583c11d462cb69fbc9cf1b49c09ce30998d5d9dace87260

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp37-cp37m-macosx_10_10_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp37-cp37m-macosx_10_10_x86_64.whl
  • Upload date:
  • Size: 22.0 MB
  • Tags: CPython 3.7m, macOS 10.10+ x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp37-cp37m-macosx_10_10_x86_64.whl
Algorithm Hash digest
SHA256 7176f9cd3f97b9045aa6d3a33abde6b2625f2505f66529198fb464aeb2ca00d3
MD5 91367063a62060c824101a1efdca4b28
BLAKE2b-256 c488fc072f23babefc46c01a760efa118063bcda7b9a1ad40b36234efd4d2e6a

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp36-cp36m-manylinux1_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp36-cp36m-manylinux1_x86_64.whl
  • Upload date:
  • Size: 15.2 MB
  • Tags: CPython 3.6m
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp36-cp36m-manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 bd7ede2bdeb0c041504918e77b2e4c5c614fc5ab33146b4977b55569da338dcd
MD5 78880523e982008eaad9d0fcff50d94d
BLAKE2b-256 6b497206bbffabef2c1d9b5bbe3c9f3c130cf00870c047537ff9ff8ec9ad99c7

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp36-cp36m-macosx_10_10_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp36-cp36m-macosx_10_10_x86_64.whl
  • Upload date:
  • Size: 22.0 MB
  • Tags: CPython 3.6m, macOS 10.10+ x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp36-cp36m-macosx_10_10_x86_64.whl
Algorithm Hash digest
SHA256 b77d1caee12e7624ec20d9357d4c134e072621da7a6e52c17c9d7e6b5e3e5664
MD5 4c383321be2dc7f86b66652193faf66c
BLAKE2b-256 5fc75002db2c3e36d462b7a0178a202cf9be88fc816a669c17ad39e71f55e2f0

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp35-cp35m-manylinux1_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp35-cp35m-manylinux1_x86_64.whl
  • Upload date:
  • Size: 15.2 MB
  • Tags: CPython 3.5m
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp35-cp35m-manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 2558705f7e01a1d5b141b98a2a73955849072d8d5ef2d03409bd8d742efd6bc5
MD5 4fa627e9fc858f9121f2c29cbd1b389c
BLAKE2b-256 ea0744070031029902fa096c8a5df6d2e76ec742d2129e3e9910eb5550507d4f

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp35-cp35m-macosx_10_10_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp35-cp35m-macosx_10_10_x86_64.whl
  • Upload date:
  • Size: 22.0 MB
  • Tags: CPython 3.5m, macOS 10.10+ x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp35-cp35m-macosx_10_10_x86_64.whl
Algorithm Hash digest
SHA256 ef5f1f390e8b10a7fe8bcb5b679e4cdad0a789c1f81beb2c9ad3f274cb44644e
MD5 e17ba42f26e79a5163f3502881856c29
BLAKE2b-256 f50c1bece1ff7dce927a1cb72dbe19cb4016c2bec54f550f0a378728157a084c

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp35-cp35m-linux_armv7l.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp35-cp35m-linux_armv7l.whl
  • Upload date:
  • Size: 18.9 MB
  • Tags: CPython 3.5m
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp35-cp35m-linux_armv7l.whl
Algorithm Hash digest
SHA256 5e945510ffbab774259ea3ab5a9c07c88fd664e08467152369782d1ebeb8e550
MD5 bdac2422b7e83b975cf7f0f05093e2a0
BLAKE2b-256 1ef797620602651338204255179ee94058b8e447767062549cf90e5a986771f1

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp34-cp34m-manylinux1_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp34-cp34m-manylinux1_x86_64.whl
  • Upload date:
  • Size: 15.2 MB
  • Tags: CPython 3.4m
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp34-cp34m-manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 84e60ad7f51674cfeb934aeeb32741266c9c23fe1a5edf7d989b09fa9d253fc6
MD5 b0116e69582275d9479324da3c6b6f1a
BLAKE2b-256 caf9b11db3823b09294d5a0bed022f84ed09b50d7229399928ae16302b7dc672

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp34-cp34m-macosx_10_10_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp34-cp34m-macosx_10_10_x86_64.whl
  • Upload date:
  • Size: 22.0 MB
  • Tags: CPython 3.4m, macOS 10.10+ x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp34-cp34m-macosx_10_10_x86_64.whl
Algorithm Hash digest
SHA256 cc86f5f997975f1f107f154c69c412da25bb0dff15d730838472755626cbe2a9
MD5 8e333dc2babf03d1a4068dd74c0b118e
BLAKE2b-256 b6696e115554efef6d8946ba410b249f98d3102c6c669271fab56b11a8d5809b

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp34-cp34m-linux_armv7l.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp34-cp34m-linux_armv7l.whl
  • Upload date:
  • Size: 18.9 MB
  • Tags: CPython 3.4m
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp34-cp34m-linux_armv7l.whl
Algorithm Hash digest
SHA256 4df4f502830f82ebb0585a68acc94c61427de22b5bd317ad45d62405cd8cd8ac
MD5 6a4686b20515442f76536d663897e51f
BLAKE2b-256 ca1a09c0c6d704bc8cc541154c525967e4f1e6d2852b39d24ed7dff1bddd9f60

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp27-cp27mu-manylinux1_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp27-cp27mu-manylinux1_x86_64.whl
  • Upload date:
  • Size: 15.2 MB
  • Tags: CPython 2.7mu
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp27-cp27mu-manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 82f627f05f5de7176a469de929cd2f5f42b6d7f173ba00ff072251fda08c8347
MD5 9728dbd4f8a91cb3cc51f8986c0d7c3a
BLAKE2b-256 668534907723e960ceddba06bc94dfb91e26eddd5580fe70a7e6553fa045bc7b

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp27-cp27mu-macosx_10_10_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp27-cp27mu-macosx_10_10_x86_64.whl
  • Upload date:
  • Size: 22.0 MB
  • Tags: CPython 2.7mu, macOS 10.10+ x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp27-cp27mu-macosx_10_10_x86_64.whl
Algorithm Hash digest
SHA256 df56c7a17e174590cfd9a65e0d0cfb8af17f191d949d0fe440435b47fc14e264
MD5 f8753e489a59fc3dcf5462f4123d4af4
BLAKE2b-256 14f1b05e5626103d220f178efc7ff724418622cebdc47745066e248ba0558da2

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp27-cp27m-manylinux1_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp27-cp27m-manylinux1_x86_64.whl
  • Upload date:
  • Size: 15.2 MB
  • Tags: CPython 2.7m
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp27-cp27m-manylinux1_x86_64.whl
Algorithm Hash digest
SHA256 9941c6804ca9da1d57adae838bed0a620f28072e39879b8f3d6d2d987faf2a30
MD5 57e0f3f504cb751a3875db35aea67d2d
BLAKE2b-256 45c526d37d3a34fc2b8d9fd559a10f14f634eb88e8c9e82c7b42d407d4dbe76d

See more details on using hashes here.

File details

Details for the file deepspeech-0.5.0a3-cp27-cp27m-macosx_10_10_x86_64.whl.

File metadata

  • Download URL: deepspeech-0.5.0a3-cp27-cp27m-macosx_10_10_x86_64.whl
  • Upload date:
  • Size: 22.0 MB
  • Tags: CPython 2.7m, macOS 10.10+ x86-64
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/1.11.0 pkginfo/1.4.2 requests/2.19.1 setuptools/39.0.1 requests-toolbelt/0.8.0 tqdm/4.26.0 CPython/3.6.2

File hashes

Hashes for deepspeech-0.5.0a3-cp27-cp27m-macosx_10_10_x86_64.whl
Algorithm Hash digest
SHA256 723daa10b494007f5605a97b725ea4b87281e6cdf2f15f67e802b11e86337efe
MD5 f2d1813602351d6484a6a59f6673a0bb
BLAKE2b-256 2022c7341cf3d7e78974b97f974a886eb3074573c723c89ce67c7c8766127aa6

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page