v1.17.0: ITREX WOQ, IPEX pipeline, extended OpenVINO export
OpenVINO
-
Enable BioGPT, Cohere, Persimmon, XGLM export by @eaidova in #709
-
Add OVModelForVision2Seq class by @eaidova in #634
from optimum.intel import OVModelForVision2Seq model = OVModelForVision2Seq.from_pretrained("nlpconnect/vit-gpt2-image-captioning", export=True) gen_tokens = model.generate(**inputs)
-
Introduce OVQuantizationConfig for NNCF quantization by @nikita-savelyevv in #638
-
Enable hybrid StableDiffusion models export via optimum-cli by @l-bat in #618
optimum-cli export openvino --model SimianLuo/LCM_Dreamshaper_v7 --task latent-consistency --dataset conceptual_captions --weight-format int8 <output_dir>
-
Convert Tokenizers by default by @apaniukov in #580
-
Custom tasks modeling by @IlyasMoutawwakil in #669
-
Add dynamic quantization config by @echarlaix in #661
from optimum.intel import OVModelForCausalLM, OVDynamicQuantizationConfig model_id = "meta-llama/Meta-Llama-3-8B" q_config = OVDynamicQuantizationConfig(bits=8, activations_group_size=32) model = OVModelForCausalLM.from_pretrained(model_id, export=True, quantization_config=q_config)
-
Transition to a newer NNCF API for PyTorch model quantization by @nikita-savelyevv in #630
ITREX
- Add ITREX weight-only quantization support by @PenghuiCheng in #455
IPEX
- Add IPEX pipeline by @jiqing-feng in #501