You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It takes a long time due to many template compilations for inference performance optimizations.
Nothing can be done to shorten it other than removing support for some GGUF file formats (which is undesirable outside of the development of llama.cpp itself), and the compilation times mainly depend on your hardware.
It will only get slower over time due to increasing support of new features and model architectures in llama.cpp.
I recommend you to switch to version 3 beta, which ships with prebuilt binaries that you can use without compiling anything.
I ran
npx --no node-llama-cpp download --cuda
and it takes a seriously long time to compile, seemingly because it's only running on one thread.Is there anything I can do to speed it up?
The text was updated successfully, but these errors were encountered: