Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Using CUDA #35

Open
ItsMeJacky opened this issue Aug 20, 2024 · 5 comments
Open

Using CUDA #35

ItsMeJacky opened this issue Aug 20, 2024 · 5 comments

Comments

@ItsMeJacky
Copy link

How to run faster-whisper using CUDA ?

@rufinus
Copy link

rufinus commented Aug 22, 2024

cudnn libs + --device cuda

on my LXC setup i followed this article https://sluijsjes.nl/2024/05/18/coral-and-nvidia-passthrough-for-proxmox-lxc-to-install-frigate-video-surveillance-server/

you just have to include the correct libs in the container, or map damn from the host to the container.

i ended up building the cointainer myself with this dockerfile:

FROM rhasspy/wyoming-whisper:2.1.0
RUN apt update && apt install -y wget software-properties-common
RUN wget https://developer.download.nvidia.com/compute/cudnn/8.9.7.29/local_installers/12.2/cudnn-local-repo-debian11-8.9.7.29_1.0-1_amd64.deb
RUN dpkg -i cudnn-local-repo-debian11-8.9.7.29_1.0-1_amd64.deb
RUN cp /var/cudnn-local-repo-debian11-8.9.7.29/cudnn-*-keyring.gpg /usr/share/keyrings/
RUN wget https://developer.download.nvidia.com/compute/cuda/repos/debian11/x86_64/cuda-keyring_1.1-1_all.deb
RUN dpkg -i cuda-keyring_1.1-1_all.deb
RUN add-apt-repository contrib && apt --allow-releaseinfo-change update
RUN apt -y install libcudnn8 libcublas-12-0
RUN rm cudnn-local-repo-debian11-8.9.7.29_1.0-1_amd64.deb

took some time to find the right versions to make it run.

@ssamjh
Copy link

ssamjh commented Oct 9, 2024

👍 For official GPU support, would be nice to speed things up :)

@ssamjh
Copy link

ssamjh commented Oct 17, 2024

👍 for this feature. I already have a GPU in my server for other applications :)

@Mithras
Copy link

Mithras commented Oct 17, 2024

Also, is it possible to make it unload model based on some --keep-alive param?

@alienatedsec
Copy link

How to run faster-whisper using CUDA ?

Simple and works out of the box - https://github.com/linuxserver/docker-faster-whisper

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants