OpenVINO Model Server C-API (preview feature)¶
Introduction¶
This document describes OpenVINO Model Server (OVMS) C API that allows OVMS to be linked into C/C++ applications. With exceptions listed at the end of this document, all capabilities of OVMS are included in the shared library.
API Description¶
Server functionalities are encapsulated in shared library built from OVMS source. To include OVMS you need to link this library with your application and use C API defined in header file.
Calling a method to start the model serving in your application initiates the OVMS as a separate thread. Then you can schedule inference both directly from app using C API and gRPC/HTTP endpoints.
Server configuration and start¶
To start OVMS you need to create OVMS_Server
object using OVMS_ServerNew
, with set of OVMS_ServerSettings
and OVMS_ModelsSettings
that describe how the server should be configured. Once the server is started using OVMS_ServerStartFromConfigurationFile
you can schedule the inferences using OVMS_Inference
. To stop server, you must call OVMS_ServerDelete
. While the server is alive you can schedule both in process inferences as well as use gRPC API to schedule inferences from remote machine. Optionally you can also enable HTTP service. Example how to use OVMS with C/C++ application is here.
Error handling¶
Most of OVMS C API functions return OVMS_Status
object pointer indicating the success or failure. Success is indicated by nullptr (NULL). Failure is indicated by returning OVMS_Status
object. The status code can be extracted using OVMS_StatusGetCode
function and the details of error can be retrieved using OVMS_StatusGetDetails
function.
The ownership of OVMS_Status
is passed to the caller of the function. You must delete the object using OVMS_StatusDelete
.
Inference¶
To execute inference using C API you must follow steps described below.
Prepare inference request¶
Create an inference request using OVMS_InferenceRequestNew
specifying which servable name and optionally version to use. Then specify input tensors with OVMS_InferenceRequestAddInput
and set the tensor data using OVMS_InferenceRequestSetData
.
Invoke inference¶
Execute inference with OVMS using OVMS_Inference
synchronous call. During inference execution you must not modify OVMS_InferenceRequest
and bound memory buffers.
Process inference response¶
If the inference was successful, you receive OVMS_InferenceRequest
object. After processing the response, you must free the response memory by calling OVMS_InferenceResponseDelete
.
To process response, first you must check for inference error. If no error occurred, you must iterate over response outputs and parameters using OVMS_InferenceResponseGetOutputCount
and OVMS_InferenceResponseGetParameterCount
. Then you must extract details describing each output and parameter using OVMS_InferenceResponseGetOutput
and OVMS_InferenceResponseGetParameter
. Example how to use OVMS with C/C++ application is here. While in example app you have only single thread scheduling inference request you can execute multiple inferences simultaneously using different threads.
Note : After inference execution is finished you can reuse the same OVMS_InferenceRequest
by using OVMS_InferenceRequestInputRemoveData
and then setting different tensor data with OVMS_InferenceRequestSetData
.
Preview limitations¶
Launching server in single model mode is not supported. You must use configuration file.
There is no direct support for jpeg/png encoded input format through C API.
There are no server live, server ready, model ready, model metadata, metrics endpoints exposed through C API.
Inference scheduled through C API does not have metrics
ovms_requests_success
,ovms_requests_fail
andovms_request_time_us
counted.You cannot turn gRPC endpoint off, REST API endpoint is optional.
There is no API for asynchronous inference.
There is no support for stateful models.