Skip to content

v1.17.0: ITREX WOQ, IPEX pipeline, extended OpenVINO export

Compare
Choose a tag to compare
@echarlaix echarlaix released this 30 May 17:29
· 305 commits to main since this release

OpenVINO

  • Enable Orion, InternLM2 export by @eaidova in #628

  • Enable OLMo export by @eaidova in #678

  • Enable Phi3 export by @eaidova in #686

  • Enable BioGPT, Cohere, Persimmon, XGLM export by @eaidova in #709

  • Enable Aquila, InternLM, XVERSE export by @eaidova in #716

  • Add OVModelForVision2Seq class by @eaidova in #634

    from optimum.intel import OVModelForVision2Seq
    
    model = OVModelForVision2Seq.from_pretrained("nlpconnect/vit-gpt2-image-captioning", export=True)
    gen_tokens = model.generate(**inputs)
  • Introduce OVQuantizationConfig for NNCF quantization by @nikita-savelyevv in #638

  • Enable hybrid StableDiffusion models export via optimum-cli by @l-bat in #618

    optimum-cli export openvino --model SimianLuo/LCM_Dreamshaper_v7 --task latent-consistency  --dataset conceptual_captions --weight-format int8 <output_dir>
    
  • Convert Tokenizers by default by @apaniukov in #580

  • Custom tasks modeling by @IlyasMoutawwakil in #669

  • Add dynamic quantization config by @echarlaix in #661

    from optimum.intel import OVModelForCausalLM, OVDynamicQuantizationConfig
    
    model_id = "meta-llama/Meta-Llama-3-8B"
    q_config = OVDynamicQuantizationConfig(bits=8, activations_group_size=32)
    model = OVModelForCausalLM.from_pretrained(model_id, export=True, quantization_config=q_config)
  • Transition to a newer NNCF API for PyTorch model quantization by @nikita-savelyevv in #630

ITREX

IPEX