Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

initialize zp, scale loaded from HF quantizer, applying quant_config #84

Closed
wants to merge 2 commits into from

Conversation

horheynm
Copy link
Member

@horheynm horheynm commented Jun 14, 2024

When base models are loaded using HF quantizer, weights are frozen.
ZP and scale attributes do not exist. If not quantized, set it to Initialize.

Tests in HF quantizer:
neuralmagic/transformers#102

@horheynm horheynm marked this pull request as draft June 14, 2024 17:39
@markurtz
Copy link
Member

@dsikka can you confirm that this will be handled by your observer refactor and we'll be able to close this out?

@dsikka
Copy link
Contributor

dsikka commented Oct 22, 2024

Is this for a dense model? Why are zero-points and scales missing?
Please expand in the PR description + provide an example model and code snipped with this issue @horheynm

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants