Skip to content

Run an Intel GPU-accelerated Wyoming protocol speech-to-text service for your Home Asssistant in Docker

License

Notifications You must be signed in to change notification settings

tannisroot/wyoming-whisper-cpp-intel-gpu-docker

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Wyoming Whisper.cpp for Intel GPUs in Docker

Run an Intel GPU-accelerated Wyoming protocol speech-to-text service for your Home Asssistant in Docker.

Utilizes whisper.cpp and Wyoming Whisper API client.

Tested on an Intel Arc A380.

Hardware requirements

  • Intel UHD Graphics for 11th generation Intel processors or newer
  • Intel Iris Xe graphics
  • Intel Arc graphics
  • Intel Server GPU
  • Intel Data Center GPU Flex Series
  • Intel Data Center GPU Max Series

Installation

  • Open the docker-compose.yaml and change the necessary values like user and path.
  • Ensure the group_add matches your render group id.
  • Install by running docker compose up.

Usage

Add this as a service for the Wyoming Integration. When setting it up, enter your host IP (127.0.0.1 if you run the container on the same host and HA is in a container too) and port 7891.

The initial request will be relatively slow, but subsequent ones will be significantly faster.

Improving accuracy

To improve accuracy, especially of difficult, uncommon words, you may use the initial prompt option. In the docker-compose.yaml file, simply change the value of

- PROMPT=""

It can set to words that are common in your commands, words that whisper.cpp is having difficulty understanding, or some brief instructions. It seems to work best when you just give it all the areas, devices and actions that control them, without any particular structure, like so:

- PROMPT="turn off turn on close stop play run set timer cancel temperature degrees weather tv lights lamp curtains roomba thermostat AC bedroom kitchen entryway corridor living room gym"

Changing model

By default, the container is set to use large-v2, but you are free to set MODEL environment variable to any supported model.

About

Run an Intel GPU-accelerated Wyoming protocol speech-to-text service for your Home Asssistant in Docker

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published