Simple but extensible API for Speech Recognition.
Project description
Speech Recognition API
Simple but extensible API for Speech Recognition.
Installation
From pip:
pip install speech_recognition_api[all]
From git:
git clone https://github.com/asai95/speech-recognition-api.git
pip install -r requirements.txt
Usage
Simple dev server:
python -m speech_recognition_api
Gunicorn:
gunicorn "speech_recognition_api:create_app()" -k uvicorn.workers.UvicornWorker -w 1 -b 127.0.0.1:8888
Celery worker:
celery -A speech_recognition_api.extra.celery_bus worker
Huey worker:
huey_consumer speech_recognition_api.extra.huey_bus.huey
Description
This project is aimed to simplify building and deploying of applications that require Speech Recognition functionality.
It is designed to work as a microservice, so it does not handle stuff like auth and rate limits.
However, it is also designed to be extensible in 3 major areas:
- Models
- File Storages
- Message Busses
There are two types of APIs available.
Synchronous API
This API is designed for simple workloads, where the machine that runs the server is capable of running a model. You probably want to limit the payload size for these routes.
Routes:
POST /sync/v1/transcribe
Accepts an audio file. File type depends on the model that is being used.
Returns an object with transcription. Response model.
Asynchronous API
This API is designed to process files asynchronously, i.e. to create tasks and process them on separate workers. Typical client flow here is as follows:
- Create a task and receive task id
- Use this task id to periodically check if it is completed.
Routes:
POST /async/v1/transcribe
Accepts an audio file. File type depends on the model that is being used.
Returns an object with async task id. Response model.
GET /async/v1/transcribe/{task_id}
Returns an object with status and a transcription (if transcription is available). Response model.
Async API also requires a worker to run the actual work.
Configuring
Configuration is done by .env file or env variables (they take preference).
The main variables required for the API and worker to run are:
MODEL
- model class path (it will do the actual audio-to-text conversion)STORAGE
- storage class path (in Async API it will be responsible for uploading/downloading files)MESSAGE_BUS
- message bus class path (in Async API it will be responsible for sending tasks to remoted workers and getting the result back from them)
These classes will be imported only when used for the fist time.
Each class may require its own variables. Please refer to config.py of the specific module to get the config reference.
Built-in classes:
Models:
Storages:
Message Busses:
Extending
It is easy to extend the API by adding models, storages and message busses.
To do that, one can just create a class that implements an interface:
Then just add a path to the class to the config file and that's it!
I suggest to distribute new modules through PyPI, so other people could reuse them.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for speech-recognition-api-0.1.0.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5793ff28a965a8ca92f39774998f13aaed5fa6ace73cc76636c15b429248dd03 |
|
MD5 | 01a742e0cd9c675f4e731ce17d1eae24 |
|
BLAKE2b-256 | aa949ad85a648c515ec747d50237b922b33a9a94688719e1edc554145e487dab |
Hashes for speech_recognition_api-0.1.0-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | b603852771b28937b79d96f5341aea598931dfd7fc742e32b8e358fd90cc3549 |
|
MD5 | fb77a8cdd925ff2064c15c2f1b88df17 |
|
BLAKE2b-256 | 4528232112d504832a22bf580485ea9815bd2aef07d739ad47e79379bfb54a7d |