-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Top probabilities broken since llama.cpp >= b4365 #104
Comments
Thank you for the bug report! The token probabilities issue should be fixed as of commit c3daede. |
You will see a performance hit as long as |
Quick update on the issue. |
Using the latest version of mikupad, the show-on-hover top probabilities function seems broken, nothing is shown. I can reproduce with llama.cpp backend version b4365 onward, works fine until b4363. In addition, since that version, inference is roughly 20% slower (varies with model).
This commit mentioned may be the culprit and may now require special handling from the perspective of the frontend.
The text was updated successfully, but these errors were encountered: