-
Notifications
You must be signed in to change notification settings - Fork 9.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Segfault with 65B model #84
Comments
Can you provide the parameters for which you get this error? |
Basically, this fails if I increase n_ctx, beyond the default 512, which I can tell isn't fully supported. I increased the mem_size allocated by ggml by adding to ctx_size, but it still uses more memory than allocated without showing any warning/error messages. These parameters
actually cause a null dereference partway through inference:
|
related discussion #71 |
@matthew-mcallister Can you try again with the code from master (which is now using mmap to load the weights)? |
EDIT: Hold on, I might be mistaken. I haven't finished converting all the tensors yet. |
OK, this works now. Fantastic, thanks for the update! |
This is the output with
-fsanitize=address
:I had to increase
ctx_size
otherwise I got this error:Is GGML trying to use more RAM than it malloc'd?
The text was updated successfully, but these errors were encountered: