-
Notifications
You must be signed in to change notification settings - Fork 791
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature request] onnx model compression #284
Comments
I think adding something like this via an experimental opt-in feature would be a good idea. In fact, this will be beneficial for all other files. For example, Surely there's a way to skip the requirement of serving extra HTTP headers? If so, I can easily update some of the models and do some testing. For example, the Compression Streams API? |
Here are some ideas regarding
Some ideas on
In my project cdn I am using .htaccess with some hardcoded data:
|
Thanks for releasing models with reduced size https://twitter.com/xenovacom/status/1698742891118493905 .
I was thinking of further reduction using compression algorithm like brotli. I have tested current whisper-base.en (~51MB) can be reduced to ~26MB using:
However huggingface_hub doesnt seem to be capable ATM huggingface/huggingface_hub#1446 .
So my idea / feature request is, whether instead of compressing on flight, it would be possible to:
In order for this to be transparent
fetch()
for browser, such .br files should be served with extra http headers:And preferably additional information on the about the original filesize (so js
fetch()
can recognise and report progress properly) i.e.The text was updated successfully, but these errors were encountered: