You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Nov 8, 2022. It is now read-only.
Hello, I read the Q8Bert paper and have tried to reproduce the experiment results.
But, on some GLUE tasks ( e.g cola, mrpc ), the differences between the fp32 results and quantized ones are much larger than the differences reported in the paper.
I tried sweeping initial learning rate but still the result was still far from the reported results.
So, I want to ask you if the experiment on Q8bert was done with default parameters set inside nlp-architect code as below.
If not, could you tell me the experiment setting.
The text was updated successfully, but these errors were encountered:
Hello, I read the Q8Bert paper and have tried to reproduce the experiment results.
But, on some GLUE tasks ( e.g cola, mrpc ), the differences between the fp32 results and quantized ones are much larger than the differences reported in the paper.
I tried sweeping initial learning rate but still the result was still far from the reported results.
So, I want to ask you if the experiment on Q8bert was done with default parameters set inside nlp-architect code as below.
If not, could you tell me the experiment setting.
The text was updated successfully, but these errors were encountered: