You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As the NNCF document says, when quantizing a Transformer model, the parameter model_type should be set to TRANSFORMER.
In my tests, if it is not set, the performance of the INT8 model will decline. If it is set, there will be a 10% performance improvement.
Describe the bug
INT8 Performance
As the NNCF document says, when quantizing a Transformer model, the parameter model_type should be set to TRANSFORMER.
In my tests, if it is not set, the performance of the INT8 model will decline. If it is set, there will be a 10% performance improvement.
INT8 Accuracy
Dataset
MVTec
Model
Other (please specify in the field below)
Steps to reproduce the behavior
OS information
OS information:
Expected behavior
works for the INT8 Model
Screenshots
No response
Pip/GitHub
GitHub
What version/branch did you use?
No response
Configuration YAML
None
Logs
Code of Conduct
The text was updated successfully, but these errors were encountered: