-
Notifications
You must be signed in to change notification settings - Fork 29
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
convert to guff and ollama #10
Comments
Same issue. |
Working on it, managed to get a gguf working |
thank you, i want to know how did you solve this problem please? |
I used the gguf llamacpp conversion tool but it was missing files so I literally copy pasted files from llama 3.1 (non instruct model) and modified the end of sentence token in generation config file |
thank you~ |
I want to deploy it via ollama, so I firstly convert it to .guff file by llama.cpp's convert_hf_to_guff.py,but I got an error that KeyError "<|user|>",so I found it not in added_tokens_decoder of tokenizer_config.json while it exists in convert_hf_to_guff.py.So I comment this line out, code can finish normally.And ollama create model also successfully completed.But when ollama run model,I got tensor number errors,
I wonder what happens after removing the <|user|> keyword and how should I fix this?
The text was updated successfully, but these errors were encountered: