Skip to main content

server for mozilla deepspeech

Project description

# DeepSpeech Server

This is an http server that can be used to test the mozilla DeepSpeech project.
You need an environment with DeepSpeech and a model to run this server.

## Installation

You first need to install deepspeech. Depending on your system you can use the
CPU package:

pip3 install deepspeech

Or the GPU package:

pip3 install deepspeech-gpu

Then you can install the deepspeech server:

python3 setup.py install

The server is also available on pypi, so you can install it with pip:

pip3 install deepspeech-server

Note that python 3.5 is the minimum version required to run the server.

## Starting the server

deepspeech-server --config config.json

You can use deepspeech without training a model yourself. Pre-trained
models are provided by Mozilla in the release page of the project (See the
download section at the bottom):

https://github.com/mozilla/DeepSpeech/releases

### Server configuration

The configuration is done with a json file, provided with the "--config" argument.
Its structure is the following one:

{
"deepspeech": {
"model" :"model.pb",
"alphabet": "alphabet.txt",
"lm": "lm.binary",
"trie": "trie"
},
"server": {
"http": {
"request_max_size": 1048576
}
}
}

The configuration file contains several sections and sub-sections.

Section "deepspeech" contains configuration of the deepspeech engine:

__model__ is the protobuf model that was generated by deepspeech

__alphabet__ is the alphabet dictionary (as available in the "data" directory of
the DeepSpeech sources).

__lm__ is the language model.

__trie__ is the trie file.

Section "server" contains configuration of the access part, with on subsection per protocol:

http configuration:

__request_max_size__ (default value: 1048576, i.e. 1MiB) is the maximum payload
size allowed by the server. A received payload size above this threshold will
return a "413: Request Entity Too Large" error.

__host__ (default value: "0.0.0.0") is the listen address of the http server.

__port__ (default value: 8080) is the listening port of the http server.


## Using the server

Inference on the model is done via http post requests. For example with the
following curl command:

curl -X POST --data-binary @[myfile.wav] http://localhost:8000/stt

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

deepspeech-server-0.4.0.zip (12.0 kB view details)

Uploaded Source

File details

Details for the file deepspeech-server-0.4.0.zip.

File metadata

File hashes

Hashes for deepspeech-server-0.4.0.zip
Algorithm Hash digest
SHA256 1352a5c04a3b0f1533424a8d4aaed5aecb585a9467521a28e9fd242c49cf0527
MD5 e4a2e15fb38b195ec500ac0cad7eaef5
BLAKE2b-256 0d0239370ce1f6e03ab3ea63b572883b6405786936eb07c2ee6b4d136eb8a972

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page