Wyoming protocol server for onnx asr speech to text system
Find a file
2025-10-31 19:04:16 +00:00
.github/workflows Update docker-publish-gpu-tensorrt.yml 2025-07-07 07:47:28 +01:00
tests Change tests 2025-10-17 21:34:57 +01:00
tools Fix tests 2025-07-14 07:51:20 +01:00
wyoming_onnx_asr Remove EN from multilingual list 2025-07-17 20:03:27 +01:00
.dockerignore General tidy 2025-07-04 08:30:46 +01:00
.gitignore Add tensorrt 2025-07-06 12:21:54 +01:00
CHANGELOG.md Fix default port in dockerfile 2025-08-02 07:27:38 +01:00
compose.gpu-trt.yaml Update changelog + other bits 2025-07-13 21:46:21 +01:00
compose.gpu.yaml Update changelog + other bits 2025-07-13 21:46:21 +01:00
compose.yaml Update changelog + other bits 2025-07-13 21:46:21 +01:00
Dockerfile Fix default port in dockerfile 2025-08-02 07:27:38 +01:00
gpu-tensorrt.Dockerfile Fix default port in dockerfile 2025-08-02 07:27:38 +01:00
gpu.Dockerfile Fix default port in dockerfile 2025-08-02 07:27:38 +01:00
LICENSE.md General tidy 2025-07-04 08:30:46 +01:00
mise.toml Add tensorrt 2025-07-06 12:21:54 +01:00
mypy.ini Initial commit 2023-10-02 10:49:18 -05:00
pyproject.toml Fix version number 2025-10-31 19:04:16 +00:00
README.md Update changelog + other bits 2025-07-13 21:46:21 +01:00
uv.lock Update onnx-asr to support parakeet v3 2025-10-17 21:28:04 +01:00

Wyoming Onnx ASR

Wyoming protocol server for the onnx-asr speech to text system.

Docker Image

docker run -it -p 10300:10300 -v /path/to/local/data:/data ghcr.io/tboby/wyoming-onnx-asr

or for gpu

docker run -it -p 10300:10300 --gpus all -v /path/to/local/data:/data ghcr.io/tboby/wyoming-onnx-asr-gpu

There is also gpu TensorRT support, but it's a huge image and doesn't seem to make much performance difference. You might want to mount in a cache folder if using it (/cache).

Local Install

Install uv

Clone the repository and use uv:

git clone https://github.com/tboby/wyoming-onnx-asr.git
cd wyoming-onnx-asr
uv sync

Run a server anyone can connect to:

uv run --uri 'tcp://0.0.0.0:10300'

The --model-en or --model-multilingual can also be a HuggingFace model but see onnx-asr for details

NOTE: Models are downloaded temporarily to the HF_HUB_CACHE directory, which defaults to ~/.cache/huggingface/hub. You may need to adjust this environment variable when using a read-only root filesystem (e.g., HF_HUB_CACHE=/tmp).

Configuration

  • Quantization: the parakeet model supports int8, but make sure to compare as performance may or may not improve.

Running tooling

Install mise and use mise run to get a list of tasks to test, format, lint, run.