composite_demo如何运行int4量化后的模型呢? #671
-
我量化后的模型运行main.py提示需要其他的文件,我下载完配置文件后提示[model_proto->ParseFromArray(serialized.data(), serialized.size())] |
Beta Was this translation helpful? Give feedback.
Answered by
zRzRzRzRzRzRzR
Jan 2, 2024
Replies: 1 comment 1 reply
-
你可以在client.py文件的141和155行里加上quantize(4)试试:
不过我用微调后的8位量化模型时报错,官方回复说是不支持量化的。 |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
是的,这个quantize(4)可以支持原生模型的在线量化,但是我们没有推出离线量化