You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When I run the compiled CUDA bitonic sorter example (linked in the README) I get this error:
Failed to launch kernels (error code an illegal memory access was encountered)!
To Reproduce
Steps to reproduce the behavior:
bend gen-cu sorter.bend > sorter.cu
nvcc sorter.cu -o sorter
prime-run ./sorter (Launches it on the GPU for Arch Linux.)
Error recieved.
Expected behavior
The program runs on the GPU.
Desktop (please complete the following information):
OS: Linux (Arch 6.9.1-arch1-1)
CPU: Intel i7-11800H
GPU: RTX 3050 Ti Mobile
GPU Driver: Nvidia open kernel modules v550.78
CUDA release 12.4, V12.4.131
Additional context
The program runs using the C codegen backend, but with the CUDA backend, it seems to fail regardless of what I do. If anyone is curious about the prime-run command, it's really just a script that forces the dGPU to handle a task - nothing fancy.
The text was updated successfully, but these errors were encountered:
The discussion is here on the hvm github. Re-posting because, hvm V2.0.13 is required and this is not the version of the hvm github, so this fix is specific to bend. (github V2.0.14 does not work at all with my bend)
I had the same issue. I cloned, HVM changed LNet seeting according to #283 , but the current repo V2.0.14 does not work with bend, and I do not know where V2.0.13 (for bend) is.
I never used cargo so excuse me if I am doing some black magic here, but this is how I fixed it for bend:
Description
When I run the compiled CUDA bitonic sorter example (linked in the README) I get this error:
To Reproduce
Steps to reproduce the behavior:
bend gen-cu sorter.bend > sorter.cu
nvcc sorter.cu -o sorter
prime-run ./sorter
(Launches it on the GPU for Arch Linux.)Expected behavior
The program runs on the GPU.
Desktop (please complete the following information):
Additional context
The program runs using the C codegen backend, but with the CUDA backend, it seems to fail regardless of what I do. If anyone is curious about the
prime-run
command, it's really just a script that forces the dGPU to handle a task - nothing fancy.The text was updated successfully, but these errors were encountered: