Replies: 1 comment 2 replies
-
Yeah I think the Quickstart guide needs another revamp into an actual guide instead of a pile of different setup options. For now I would recommend using text-generation-webui as the backend along with the addon. I should be able to provide some better instructions in the next few days. |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I followed instructions for installing the Addon, hoping it contained everything that would be in the manual instructions. I'm trying to run on Home Assistant Green, which should be at least as powerful as a Pi, but still no GPU so
cpp
models are probably the only realistic choice.When I loaded the addon, the Web UI shows up, but there are no models. So I found the manual install instructions to load one from the
/dist
folder in this repo (though the docs incorrectly link to/docs/dist
, probably a relative URL path mistake) which I have to download and add toaddon_configs
(why not include it in the addon?).But then I still don't have it in the Web UI, that needs models in a different folder (though putting the model there and not knowing what settings to use, there was errors and a stack trace in the web ui when attempting to load it).
So I guess I have to set up the integration as well, the custom component from HACS. There I get the "Select Backend" prompt and based on the docs, of the 6 options I'd just choose "Llama.cpp (HuggingFace)". Since it's a local file I'd think "Llama.cpp (local model)" would make more sense (since I have it locally, and didn't get it from HuggingFace), but when typing in the name of the model file there, it can't find it.
After that I'm prompted "please configure llama.cpp for the model" with a bunch of options about quantization. I assume this information needs to be correctly set to the specifics of the downloaded file? Why can't it detect it automatically? I have no idea what to put here but the field "HuggingFace Model" is set to
TheBloke/phi-2-GGUF
which leads me to believe this will download a remote model from huggingface, not use the local one - though the docs say to go this path by default.No matter which Quantization I choose, however, upon submitting it hangs for a long while and then "Unknown Error Occured".
I'm also thinking that this model loading scheme wouldn't use the addon at all, making the addon pointless, but that's not how the docs guided me, so idk. The component setup path for using text-generation-webui says "for a remote instance" in the docs, as in, specifically not a local one, so I guess I wouldn't use that. But I tried it, and couldn't get past the config screen where I needed a bunch of things the addon doesn't mention, like an OpenAI Api key and the model name.
When I tried loading the provided model in the addon web ui, and used the "llama.cpp" model loader, I got:
What's the simplest possible setup to get this thing running?
Beta Was this translation helpful? Give feedback.
All reactions