Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Suggestion for Official Releases of LLMs: Include Quantized Versions #17

Open
epochaudio opened this issue May 28, 2024 · 1 comment
Open
Labels
wontfix This will not be worked on

Comments

@epochaudio
Copy link

对于官方发布的 LLMS 大模型,建议在未来可以附上 awq 和 gptq 的量化版本。这种做法几乎没有成本,但却能帮助许多缺乏 GPU 的潜在用户。这会让用户在使用模型时更加方便,因为大家普遍认为官方发布的量化版本更具权威性。
For officially released LLMs, it is suggested that awq and gptq quantized versions be included in the future. This practice incurs almost no cost but could benefit many potential users who lack GPUs. It would also be more convenient for users as official quantized versions are generally considered more authoritative.

@objecti0n
Copy link
Collaborator

Please use the community's version first: https://huggingface.co/legraphista/internlm2-math-plus-1_8b-IMat-GGUF
We will consider adding quantized versions after.

@objecti0n objecti0n added the wontfix This will not be worked on label Jun 6, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
wontfix This will not be worked on
Projects
None yet
Development

No branches or pull requests

2 participants