Skip to content

Latest commit

 

History

History
20 lines (12 loc) · 1002 Bytes

File metadata and controls

20 lines (12 loc) · 1002 Bytes

Introduction to OpenVINO™ Model Server

This notebook demonstrates how to deploy a model server and request predictions from a client application.

OpenVINO Model Server (OVMS) is a high-performance system for serving models. Implemented in C++ for scalability and optimized for deployment on Intel architectures, the model server uses the same architecture and API as TensorFlow Serving and KServe while applying OpenVINO for inference execution. Inference service is provided via gRPC or REST API, making deploying new algorithms and AI experiments easy.

ovms_high_level

Notebook Contents

The notebook covers following steps:

  • Prepare Docker
  • Preparing a Model Repository
  • Start the Model Server Container
  • Prepare the Example Client Components

Installation Instructions

If you have not installed all required dependencies, follow the Installation Guide.