Skip to content

Releases: huggingface/optimum-intel

v1.19.0

10 Sep 21:57
Compare
Choose a tag to compare

v1.18.3: Patch release

19 Aug 09:16
Compare
Choose a tag to compare

v1.18.2: Patch release

06 Aug 16:13
Compare
Choose a tag to compare

Full Changelog: v1.18.1...v1.18.2

v1.18.1: Patch release

09 Jul 16:13
Compare
Choose a tag to compare

Full Changelog: v1.18.0...v1.18.1

v1.18.0: Arctic, Jais, OpenVINO pipelines

26 Jun 23:21
Compare
Choose a tag to compare

OpenVINO

from optimum.intel.pipelines import pipeline

# Load openvino model
ov_pipe = pipeline("text-generation", "helenai/gpt2-ov", accelerator="openvino")
# Load pytorch model and convert it to openvino before inference
pipe = pipeline("text-generation", "gpt2", accelerator="openvino")

IPEX

v1.17.2: Patch release

07 Jun 19:14
Compare
Choose a tag to compare

v1.17.1: Patch release

06 Jun 15:40
Compare
Choose a tag to compare

v1.17.0: ITREX WOQ, IPEX pipeline, extended OpenVINO export

30 May 17:29
Compare
Choose a tag to compare

OpenVINO

  • Enable Orion, InternLM2 export by @eaidova in #628

  • Enable OLMo export by @eaidova in #678

  • Enable Phi3 export by @eaidova in #686

  • Enable BioGPT, Cohere, Persimmon, XGLM export by @eaidova in #709

  • Enable Aquila, InternLM, XVERSE export by @eaidova in #716

  • Add OVModelForVision2Seq class by @eaidova in #634

    from optimum.intel import OVModelForVision2Seq
    
    model = OVModelForVision2Seq.from_pretrained("nlpconnect/vit-gpt2-image-captioning", export=True)
    gen_tokens = model.generate(**inputs)
  • Introduce OVQuantizationConfig for NNCF quantization by @nikita-savelyevv in #638

  • Enable hybrid StableDiffusion models export via optimum-cli by @l-bat in #618

    optimum-cli export openvino --model SimianLuo/LCM_Dreamshaper_v7 --task latent-consistency  --dataset conceptual_captions --weight-format int8 <output_dir>
    
  • Convert Tokenizers by default by @apaniukov in #580

  • Custom tasks modeling by @IlyasMoutawwakil in #669

  • Add dynamic quantization config by @echarlaix in #661

    from optimum.intel import OVModelForCausalLM, OVDynamicQuantizationConfig
    
    model_id = "meta-llama/Meta-Llama-3-8B"
    q_config = OVDynamicQuantizationConfig(bits=8, activations_group_size=32)
    model = OVModelForCausalLM.from_pretrained(model_id, export=True, quantization_config=q_config)
  • Transition to a newer NNCF API for PyTorch model quantization by @nikita-savelyevv in #630

ITREX

IPEX

v1.16.1: Patch release

25 Apr 08:09
Compare
Choose a tag to compare

v1.16.0: OpenVINO config, SD hybrid quantization

25 Mar 11:56
Compare
Choose a tag to compare

Add hybrid quantization for Stable Diffusion pipelines by @l-bat in #584

from optimum.intel import OVStableDiffusionPipeline, OVWeightQuantizationConfig

model_id = "echarlaix/stable-diffusion-v1-5-openvino"
quantization_config = OVWeightQuantizationConfig(bits=8, dataset="conceptual_captions")
model = OVStableDiffusionPipeline.from_pretrained(model_id, quantization_config=quantization_config)

Add openvino export configs by @eaidova in #568

Enabling OpenVINO export for the following architectures enabled : Mixtral, ChatGLM, Baichuan, MiniCPM, Qwen, Qwen2, StableLM

Add support for export and inference for StarCoder2 models by @eaidova in #619