How do I configure llama.cpp to use my iGPU instead of the GPU? #12443
-
I have an iGPU and a GPU: I would like to run 1 instance of llama.cpp using my GPU - AMD Radeon™ RX 7600 XT (RADV NAVI33) (this is currently working fine). And another instance on a different port using the iGPU - AMD Radeon Graphics (RADV GFX1103_R1). I've tried multiple settings (for MESA_VK_DEVICE_SELECT and for LLAMA_ARG_MAIN_GPU) but llama.cpp always pick the GPU instead of the iGPU: Any help would be really appreciated! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
You can set the Vulkan device(s) to be used with |
Beta Was this translation helpful? Give feedback.
You can set the Vulkan device(s) to be used with
GGML_VK_VISIBLE_DEVICES
in a similar way to how it works with CUDA. In your case you would useGGML_VK_VISIBLE_DEVICES=0
for your iGPU,GGML_VK_VISIBLE_DEVICES=1
for your dGPU (which is also what it defaults to), orGGML_VK_VISIBLE_DEVICES=0,1
or even1,0
for both.