v1.5.0: OpenVINO quantization
Quantization
- Add
OVQuantizer
enabling OpenVINO NNCF post-training static quantization (#50) - Add
OVTrainer
enabling OpenVINO NNCF quantization aware training (#67) - Add
OVConfig
the configuration which contains the quantization process informations (#65)
The quantized model resulting from the OVQuantizer
and the OVTrainer
are exported to the OpenVINO IR and can be loaded with the corresponding OVModelForXxx
to perform inference with OpenVINO Runtime.
OVModel
Add OVModelForCausalLM
enabling OpenVINO Runtime for models with a causal language modeling head (#76)