Best backend & model for discrete GPU on Raspberry Pi? #243
-
I have a Raspberry Pi with a discrete AMD GPU: https://github.com/pepijndevos/rpi-cm5io-mini-itx AFAIK LM Studio doesn't support ARM Linux and Ollama doesn't support Vulkan. It seems like you can directly interface with ollama.cpp, which does support Vulkan on ARM, although I'll have to compile it myself with GPU support I guess. What is less clear to me is how tool use or function calling works in home-llm and if I need to use your specific models for that. |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
llama-cpp-python doesn't seem to actually use my GPU: abetlen/llama-cpp-python#1826 I tried using the upstream llama.cpp server with the generic openai API. |
Beta Was this translation helpful? Give feedback.
-
With Home 3B I get the following logs:
Actually it seems like it doesn't support this type of device? I added an actual switch device and it works! |
Beta Was this translation helpful? Give feedback.
With Home 3B I get the following logs: