-
Notifications
You must be signed in to change notification settings - Fork 879
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Certain characters crash bitnet model inference? #102
Comments
I tried your characters using a Docker container and didn't encounter any errors.
|
@kth8 Thanks for testing this out, I notice a difference of I re-ran the setup script, and now at least I get the following error message:
I'm downloading the other 2 reference 1-bit models and will test the same issue with them. |
Code is in my repo if you want to take a look: https://github.com/kth8/bitnet |
The model build readme doesn't list tl2 as an option for quant_type
Then running the command: Results in:
Where as trying out the model "bitnet_b1_58-3B\ggml-model-i2_s.gguf" results in:
So whilst it works, it outputs � instead of £. @kth8 Your linked huggingface instructions state to run inference using the llama-cli via In your dockerfile you recursively git clone this repo, so you're also using the 3rdparty forked repo instead of the latest llama.cpp as instructed in your huggingface model readme. So, perhaps it's an issue introduced whilst building the GGUF file (via setup_env.py) using the 3rdparty repo instead of an inference issue as your docker file doesn't run into the inference issue? |
The Hugging Face repo is not mine. I just linked to it as reference for where I got the GGUF model file. |
I've been trying out BitNet on debian via WSL2 & Docker, I can confirm that this issue does not occur on that environment, it must be a windows only issue? Here's my working environment: #110 |
I've been working on securing the user input, escaping invalid characters, however I've encountered a few prompts which cause the llama-cli to abruptly halt:
The command abruptly halts at system_info, offering no further logs.
This also occurs for
--prompt "¬"
and some other advanced unicode characters.Where as it works for normal characters:
Is this because the unicode tokens haven't been tokenized in the model? Or is there a llamma unicode fix for this issue?
I could probably filter out the incompatible characters from the user prompt input if this cannot be worked around, is there a list of incompatible characters?
Thanks
The text was updated successfully, but these errors were encountered: