Releases: huggingface/optimum-intel
Releases · huggingface/optimum-intel
v1.11.1: Patch release
- Fix compatibility with
optimum
by @echarlaix in b4663b4
Full Changelog: v1.11.0...v1.11.1
v1.11.0: MPT, TIMM models, VAE image processor
OpenVINO
- Fix SDXL model U-NET component static reshaping by @eaidova in #390
- Allow changing pkv precision by @AlexKoff88 in #393
- Removed pkv history from quantization statistics of decoders by @AlexKoff88 in #394
- Add audio tasks for OpenVINO inference by @helena-intel in #396
- Do not download ONNX model in SD pipeline if not needed by @eaidova in #402
- Enable loading of Text Inversion at runtime for OpenVINO SD pipelines by @sammysun0711 in #400
- Enable Timm models OpenVINO export and inference @sawradip in #404
- Fix OpenVINO Timm models loading by @echarlaix in #413
- Add VAE image processor by @echarlaix in #421
- Enable MPT OpenVINO export and inference by @echarlaix in #425
Neural Compressor
- Fixed ONNX export for
neural-compressor>=2.2.2
by @PenghuiCheng in #409 - Enable ONNX export for INC PTQ model by @echarlaix in #373
- Fix INC CLI by @echarlaix in #426
Full Changelog: https://github.com/huggingface/optimum-intel/commits/v1.11.0
v1.10.1: Patch release
- Set minimum
optimum
version by @echarlaix in #382 - Fix compilation step so that it can be performed before inference by @echarlaix in #384
v1.10.0: Stable Diffusion XL pipelines
Stable Diffusion XL
Enable SD XL OpenVINO export and inference for text-to-image and image-to-image tasks by @echarlaix in #377
from optimum.intel import OVStableDiffusionXLPipeline
model_id = "stabilityai/stable-diffusion-xl-base-0.9"
pipeline = OVStableDiffusionXLPipeline.from_pretrained(model_id, export=True)
prompt = "sailing ship in storm by Leonardo da Vinci"
image = pipeline(prompt).images[0]
pipeline.save_pretrained("openvino-sd-xl-base-0.9")
More examples in documentation
Full Changelog: v1.9.0...v1.10.0
v1.9.4: Patch release
- Fix
OVDataLoader
for NNCF quantization aware training fortransformers
> v4.31.0 by @echarlaix in #376
Full Changelog: v1.9.3...v1.9.4
v1.9.3: Patch release
- Improved performance of decoders by @AlexKoff88 #354
- Fix openvino model integration compatibility for optimum > v1.9.0 by @echarlaix in #365
Full Changelog: v1.9.2...v1.9.3
v1.9.2: Patch release
- Fix INC distillation to be compatible with
neural-compressor
v2.2.0 breaking changes by @echarlaix in #338
v1.9.1: Patch release
- Fix inference for OpenVINO export for causal language models by @echarlaix in #351
v1.9.0: OpenVINO models improvements, TorchScript export, INC quantized SD pipeline
OpenVINO and NNCF
- Ensure compatibility for OpenVINO
v2023.0
by @jiwaszki in #265 - Add Stable Diffusion quantization example by @AlexKoff88 in #294 #304 #326
- Enable decoder quantized models export to leverage cache by @echarlaix in #303
- Set height and width during inference for static models Stable Diffusion models by @echarlaix in #308
- Set batch size to 1 by default for Wav2Vec2 for NNCF compatibility
v2.5.0
@ljaljushkin in #312 - Ensure compatibility for NNCF
v2.5
by @ljaljushkin in #314 - Fix OVModel for BLOOM architecture by @echarlaix in #340
- Add SD OV model height and width attribute and fix export for
torch>=v2.0.0
by @eaidova in #342
Intel Neural Compressor
- Add
TSModelForCausalLM
to enable TorchScript export, loading and inference for causal lm models by @echarlaix in #283 - Remove INC deprecated classes by @echarlaix in #293
- Enable IPEX model inference for text generation task by @jiqing-feng in #227 #300
- Add
INCStableDiffusionPipeline
to enable INC quantized Stable Diffusion model loading by @echarlaix in #305 - Enable the possibility to provide a quantization function and not a calibration dataset during INC static PTQ by @PenghuiCheng in #309
- Fix
INCSeq2SeqTrainer
evaluation step by @AbhishekSalian in #335 - Fix
INCSeq2SeqTrainer
padding step by @echarlaix in #336
Full Changelog: https://github.com/huggingface/optimum-intel/commits/v1.9.0
v1.8.1: Patch release
- Fix OpenVINO Trainer for transformers >= v4.29.0 by @echarlaix in #328
Full Changelog: v1.8.0...v1.8.1