0.13.0 🌋
This version introduces experimental GPU support using Vulkan. While Vulkan integration is still a work in progress - especially regarding shader performance - this marks the first step toward full GPU support.
How to build Distributed Llama with Vulkan support?
DLLAMA_VULKAN=1 make dllama
To run Distributed Llama with Vulkan please add the --gpu-index 0
argument. For example:
./dllama inference ... --gpu-index 0
Please ensure that the Vulkan SDK is installed on your machine. You can run the following command to check:
vulkaninfo