-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: Switch vulkan images to the kompute backend #83
base: main
Are you sure you want to change the base?
Conversation
dnf install -y git cmake ninja-build gcc gcc-c++ vulkan-loader-devel vulkan-tools && \ | ||
dnf install -y mesa-vulkan-drivers-24.1.2-101.el9.aarch64 && \ | ||
dnf versionlock mesa-vulkan-drivers-24.1.2-101.el9.aarch64 && \ | ||
dnf install -y git cmake ninja-build gcc gcc-c++ vulkan-loader-devel vulkan-tools fmt-devel && \ | ||
dnf copr enable -y jeffmaury/shaderc epel-9-aarch64 && \ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This part was added recently because it was required by the llama-cpp Vulkan build, is it still required for kompute ?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, I updated the mesa package because el9 has switched to a newer llvm.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sorry I was not clear enough. I was mentioning the sharerc/glslc part
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah, yes, kompute also needs glslc
Signed-off-by: Sergio Lopez <[email protected]>
The latest versions of MoltenVK (tested with 1.2.11) are not able to deal properly with llama.cpp's Vulkan (GGML_VULKAN) backend. Switch our Vulkan images to use the Kompute (GGML_KOMPUTE) backend instead, which works fine with MoltenVK and will eventually achieve the same performance as Metal on macOS bare-metal. Signed-off-by: Sergio Lopez <[email protected]>
665759f
to
188866a
Compare
The latest versions of MoltenVK (tested with 1.2.11) are not able to deal properly with llama.cpp's Vulkan (GGML_VULKAN) backend.
Switch our Vulkan images to use the Kompute (GGML_KOMPUTE) backend instead, which works fine with MoltenVK and will eventually achieve the same performance as Metal on macOS bare-metal.