v1.18.0: Arctic, Jais, OpenVINO pipelines
OpenVINO
- Enable Arctic, Jais export by @eaidova in #726
- Enable GLM-4 export by @eaidova in #776
- Move data-driven quantization after model export for text-generation models by @nikita-savelyevv in #721
- Create default token_type_ids when needed for inference by @echarlaix #757
- Resolve default int4 config for local models by @eaidova in #760
- Update to NNCF 2.11 by @nikita-savelyevv in #763
- Fix quantization config by @echarlaix in #773
- Expose trust remote code argument when generating calibration dataset for datasets >= v2.20.0 by @echarlaix #767
- Add pipelines by @echarlaix in #740
from optimum.intel.pipelines import pipeline
# Load openvino model
ov_pipe = pipeline("text-generation", "helenai/gpt2-ov", accelerator="openvino")
# Load pytorch model and convert it to openvino before inference
pipe = pipeline("text-generation", "gpt2", accelerator="openvino")
IPEX
- Enable IPEX patching for llama for >= v2.3 by @jiqing-feng in #725
- Refactor llama modeling for IPEX patching by @faaany in #728
- Refactor model loading by @jiqing-feng in #752