Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: Switch vulkan images to the kompute backend #83

Open
wants to merge 2 commits into
base: main
Choose a base branch
from

Conversation

slp
Copy link

@slp slp commented Nov 28, 2024

The latest versions of MoltenVK (tested with 1.2.11) are not able to deal properly with llama.cpp's Vulkan (GGML_VULKAN) backend.

Switch our Vulkan images to use the Kompute (GGML_KOMPUTE) backend instead, which works fine with MoltenVK and will eventually achieve the same performance as Metal on macOS bare-metal.

@slp slp requested a review from a team as a code owner November 28, 2024 17:44
@slp slp requested review from dgolovin, jeffmaury and cdrage November 28, 2024 17:44
@jeffmaury jeffmaury changed the title Switch vulkan images to the kompute backend feat: Switch vulkan images to the kompute backend Nov 29, 2024
dnf install -y git cmake ninja-build gcc gcc-c++ vulkan-loader-devel vulkan-tools && \
dnf install -y mesa-vulkan-drivers-24.1.2-101.el9.aarch64 && \
dnf versionlock mesa-vulkan-drivers-24.1.2-101.el9.aarch64 && \
dnf install -y git cmake ninja-build gcc gcc-c++ vulkan-loader-devel vulkan-tools fmt-devel && \
dnf copr enable -y jeffmaury/shaderc epel-9-aarch64 && \
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This part was added recently because it was required by the llama-cpp Vulkan build, is it still required for kompute ?

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, I updated the mesa package because el9 has switched to a newer llvm.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sorry I was not clear enough. I was mentioning the sharerc/glslc part

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ah, yes, kompute also needs glslc

slp added 2 commits December 11, 2024 16:44
The latest versions of MoltenVK (tested with 1.2.11) are not able
to deal properly with llama.cpp's Vulkan (GGML_VULKAN) backend.

Switch our Vulkan images to use the Kompute (GGML_KOMPUTE) backend
instead, which works fine with MoltenVK and will eventually achieve
the same performance as Metal on macOS bare-metal.

Signed-off-by: Sergio Lopez <[email protected]>
@slp slp force-pushed the switch-to-kompute branch 3 times, most recently from 665759f to 188866a Compare December 13, 2024 18:32
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants