Write a Client Application#

OpenVINO™ Model Server supports multiple APIs, for easy integration with systems using one of them for inference. The APIs are:

  • one compatible with TensorFlow Serving,

  • KServe API for inference

  • OpenAI API for text generation.

  • OpenAI API for embeddings

  • Cohere API for reranking

Both TFS and KServe APIs work on gRPC and REST interfaces. The REST API endpoints for generative use cases support both streamed and unary responses.

Check the following articles to learn more about the supported APIs:

In this section you can find short code samples to interact with OpenVINO Model Server endpoints via: