-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug: New DeepSeek-R1-Distill-Qwen models do not load #684
Comments
Yep not loading any DeepSeek Distill -- vocabulary: unknown pre-tokenizer type: 'deepseek-r1 |
Serg Gini — Yesterday at 3:19 AM most probably llamafile just need update on master from llama.cpp |
I was able to build my own GGUF and get llamafile to load it. It will be in my Brad's LLMs repo "soon". Watch for it in the |
yeah thanks but I want to use my own gguf - will wait for the update - hope soon |
Also:
I need 14b or 32b and want rocm gpu offload optimization for a Linux server. May have to get complicated and start using SGLang. |
This has been fixed upstream: |
Have you been able to compile it with llamafile? |
someone needs to fork this thing.. just read (discord) the lead dev just took a job from google.. of course ; >.. don't blame her |
I've fixed this in pull request #687 |
#687 just got merged! please give it a try |
works with DeepSeek-R1-Distill-Qwen-1.5B-Q8_0.gguf for me |
I don't know how to compile from source file. |
Might I know, is there any solution Although many professionals have mentioned upgrading to a new version of llama.cpp, |
It's not in the main release yet, you will have to compile from source for now |
git clone https://github.com/Mozilla-Ocho/llamafile
cd llamafile
make -j$(nproc)
make -j$(nproc) install PREFIX=$HOME/bin/llamafile Error: install: cannot stat 'o//stable-diffusion.cpp/main': No such file or directory
make: *** [Makefile:62: install] Error 1 Also: git clone https://github.com/Mozilla-Ocho/llamafile
cd llamafile
rm -rf llama.cpp
git clone https://github.com/ggerganov/llama.cpp
make -j$(nproc) Error:
The include files appear to have been restructured for llama.cpp, making the build fail. I didn't see an easy way to update the includes path, so I revised the includes manually: With these changes in place, the build still fails. |
Contact Details
[email protected]
What happened?
When attempting to load a DeepSeek-R1-DeepSeek-Distill-Qwen-GGUF model, llamafile fails to load the model -- any of 1.5b, 7b, 14b, or 32b. This occurs using llamafiler, llamafile, or as a .llamafile conversion under traditional or --v2.
Version
llamafile v0.9.0
What operating system are you seeing the problem on?
Linux
Relevant log output
The text was updated successfully, but these errors were encountered: