-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
bug: Jan is not using GPU #3737
Comments
I'm having a similar issue. I can only use my 4090 through Vulcan, not CUDA. I also have a Tesla P40 installed, and Jan sees it but can't use it. Of course, the P40 doesn't support Vulcan. In the past, I've been able to use the P40 and 4090 simultaneously to load large models. I did have some driver issues with a recent upgrade to Win11 (BSOD every few hours). Possibly not fully resolved? Not sure what to look at or what to try. |
@louis-jan is this related to #3558 |
In my case, it happens with models I've imported manually as well. And the models load and run fine on CPU (just slower, obviously) |
Hi @chlimouj Could you please share the log file, and what NGL settings are you using? |
Same here. Log file attached.
NGL was set to default 33 layers. After that AI model was loaded into VRAM and worked. |
Hi @chlimouj Yeah, with the fix, you'll need to re-import it for it to work properly. Which means it will retrieve the correct metadata from the model when importing now. |
Yeah, we're working on bundling these dependencies into the app by default. It will increase the app size a bit, but it won't require any additional actions from the user. |
Will be fixed with #3690 |
how do you install this fucking cortex extension? |
also how to change the models folder? |
Jan version
0.5.4
Describe the Bug
Jan is not using GPU
Steps to Reproduce
Model: Bielik-11B-v2.3-Instruct.Q8_0.gguf
GPU: NVIDIA RTX 4070 SUPER
Screenshots / Logs
What is your OS?
The text was updated successfully, but these errors were encountered: