Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
fix: more robust check whether the HF model is quantized (#11891)
Removes the check of `model.is_quantized` and adds more robust way of checking for 4bit and 8bit quantization in the `huggingface_pipeline.py` script. I had to make the original change on the outdated version of `transformers`, because the models had this property before. Seems redundant now. Fixes: #11809 and #11759
- Loading branch information