Skip to main content

Simple but extensible API for Speech Recognition.

Project description

PyPI - Version PyPI - Python Version PyPI - License PyPI - Downloads codecov

Speech Recognition API

Simple but extensible API for Speech Recognition.

Installation

From pip:

pip install speech_recognition_api[all]

From git:

git clone https://github.com/asai95/speech-recognition-api.git
pip install -r requirements.txt

Usage

Simple dev server:

python -m speech_recognition_api

Gunicorn:

gunicorn "speech_recognition_api:create_app()" -k uvicorn.workers.UvicornWorker -w 1 -b 127.0.0.1:8888

Celery worker:

celery -A speech_recognition_api.extra.celery_bus worker

Huey worker:

huey_consumer speech_recognition_api.extra.huey_bus.huey

Description

This project is aimed to simplify building and deploying of applications that require Speech Recognition functionality.

It is designed to work as a microservice, so it does not handle stuff like auth and rate limits.

However, it is also designed to be extensible in 3 major areas:

  • Models
  • File Storages
  • Message Busses

There are two types of APIs available.

Synchronous API

This API is designed for simple workloads, where the machine that runs the server is capable of running a model. You probably want to limit the payload size for these routes.

Routes:

POST /sync/v1/transcribe

Accepts an audio file. File type depends on the model that is being used.

Returns an object with transcription. Response model.

Asynchronous API

This API is designed to process files asynchronously, i.e. to create tasks and process them on separate workers. Typical client flow here is as follows:

  • Create a task and receive task id
  • Use this task id to periodically check if it is completed.

Routes:

POST /async/v1/transcribe

Accepts an audio file. File type depends on the model that is being used.

Returns an object with async task id. Response model.

GET /async/v1/transcribe/{task_id}

Returns an object with status and a transcription (if transcription is available). Response model.

Async API also requires a worker to run the actual work.

Configuring

Configuration is done by .env file or env variables (they take preference).

The main variables required for the API and worker to run are:

  • MODEL - model class path (it will do the actual audio-to-text conversion)
  • STORAGE - storage class path (in Async API it will be responsible for uploading/downloading files)
  • MESSAGE_BUS - message bus class path (in Async API it will be responsible for sending tasks to remoted workers and getting the result back from them)

These classes will be imported only when used for the fist time.

Each class may require its own variables. Please refer to config.py of the specific module to get the config reference.

Built-in classes:

Models:

Storages:

Message Busses:

Extending

It is easy to extend the API by adding models, storages and message busses.

To do that, one can just create a class that implements an interface:

Then just add a path to the class to the config file and that's it!

I suggest to distribute new modules through PyPI, so other people could reuse them.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

speech-recognition-api-0.1.1.tar.gz (15.8 kB view hashes)

Uploaded Source

Built Distribution

speech_recognition_api-0.1.1-py3-none-any.whl (27.0 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page