Skip to content

Latest commit

 

History

History
 
 

117-model-server

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 

Introduction to OpenVINO™ Model Server

This notebook demonstrates how to deploy a model server and request predictions from a client application.

OpenVINO Model Server (OVMS) is a high-performance system for serving models. Implemented in C++ for scalability and optimized for deployment on Intel architectures, the model server uses the same architecture and API as TensorFlow Serving and KServe while applying OpenVINO for inference execution. Inference service is provided via gRPC or REST API, making deploying new algorithms and AI experiments easy.

ovms_high_level

Notebook Contents

The notebook covers following steps:

  • Prepare Docker
  • Preparing a Model Repository
  • Start the Model Server Container
  • Prepare the Example Client Components

Installation Instructions

If you have not installed all required dependencies, follow the Installation Guide.