GGML_ASSERT: /Users/runner/work/node-llama-cpp/node-llama-cpp/llama/llama.cpp/llama.cpp:5052: n_tokens <= n_batch
#94
-
Hi there, I'm getting
I can work around it by using
I could put together a better repro of this, but wondering, has anyone else run into this? Or does anyone have suggestions here on what to change? Thank you for any help, I appreciate it! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
There's currently an issue with prompts that are longer than the |
Beta Was this translation helpful? Give feedback.
There's currently an issue with prompts that are longer than the
batchSize
; it'll be fixed as part of #85.For a workaround for now, see #76