-
Notifications
You must be signed in to change notification settings - Fork 367
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG]: Failed to load ./runtimes/win-x64/native/cuda12/llama.dll #1014
Comments
It's finding the CUDA binaries but failing to load them for some reason, probably a prerequisite is missing. |
What kind of prerequisite should I check for? |
Th CUDA toolkit is the most common one people are missing (https://developer.nvidia.com/cuda-toolkit) |
First line of subject tells that CUDA Toolkit is installed, isn't it? Also |
Oops, missed that! Unfortunately that makes it a lot trickier to debug, something like dependency walker might help (but it's hard to use).
That log message is coming from here, it's referring to the logs you're already looking at. |
Description
Installed CUDA12 toolkit:
C:\Users\Home>nvcc --version
Installed packeges:
Run:
using LLama.Native;
NativeLibraryConfig.All
.WithCuda()
.SkipCheck(true)
.WithAutoFallback(false)
.WithLogCallback((level, message) => Console.Write($"{level}: {message}"));
NativeApi.llama_empty_call();
Outputs:
And it works fine (slow) when i use CPU backend. So what can I do to run model on GPU with CUDA?
Reproduction Steps
run example code
Environment & Configuration
Known Workarounds
No response
The text was updated successfully, but these errors were encountered: