Group Infer Request#
- group ov_infer_request_c_api
The definitions & operations about infer request.
Functions
-
ov_infer_request_set_tensor(ov_infer_request_t *infer_request, const char *tensor_name, const ov_tensor_t *tensor)#
Set an input/output tensor to infer on by the name of tensor.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
tensor_name – Name of the input or output tensor.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_set_tensor_by_port(ov_infer_request_t *infer_request, const ov_output_port_t *port, const ov_tensor_t *tensor)#
Set an input/output tensor to infer request for the port.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
port – Port of the input or output tensor, which can be got by calling ov_model_t/ov_compiled_model_t interface.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_set_tensor_by_const_port(ov_infer_request_t *infer_request, const ov_output_const_port_t *port, const ov_tensor_t *tensor)#
Set an input/output tensor to infer request for the port.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
port – Const port of the input or output tensor, which can be got by call interface from ov_model_t/ov_compiled_model_t.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_set_input_tensor_by_index(ov_infer_request_t *infer_request, const size_t idx, const ov_tensor_t *tensor)#
Set an input tensor to infer on by the index of tensor.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
idx – Index of the input port. If
idx
is greater than the number of model inputs, an error will return.tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_set_input_tensor(ov_infer_request_t *infer_request, const ov_tensor_t *tensor)#
Set an input tensor for the model with single input to infer on.
Note
If model has several inputs, an error will return.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_set_output_tensor_by_index(ov_infer_request_t *infer_request, const size_t idx, const ov_tensor_t *tensor)#
Set an output tensor to infer by the index of output tensor.
Note
Index of the output preserved accross ov_model_t, ov_compiled_model_t.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
idx – Index of the output tensor.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_set_output_tensor(ov_infer_request_t *infer_request, const ov_tensor_t *tensor)#
Set an output tensor to infer models with single output.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_get_tensor(const ov_infer_request_t *infer_request, const char *tensor_name, ov_tensor_t **tensor)#
Get an input/output tensor by the name of tensor.
Note
If model has several outputs, an error will return.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
tensor_name – Name of the input or output tensor to get.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_get_tensor_by_const_port(const ov_infer_request_t *infer_request, const ov_output_const_port_t *port, ov_tensor_t **tensor)#
Get an input/output tensor by const port.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
port – Port of the tensor to get.
port
is not found, an error will return.tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_get_tensor_by_port(const ov_infer_request_t *infer_request, const ov_output_port_t *port, ov_tensor_t **tensor)#
Get an input/output tensor by port.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
port – Port of the tensor to get.
port
is not found, an error will return.tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_get_input_tensor_by_index(const ov_infer_request_t *infer_request, const size_t idx, ov_tensor_t **tensor)#
Get an input tensor by the index of input tensor.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
idx – Index of the tensor to get.
idx
. If the tensor with the specifiedidx
is not found, an error will return.tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_get_input_tensor(const ov_infer_request_t *infer_request, ov_tensor_t **tensor)#
Get an input tensor from the model with only one input tensor.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_get_output_tensor_by_index(const ov_infer_request_t *infer_request, const size_t idx, ov_tensor_t **tensor)#
Get an output tensor by the index of output tensor.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
idx – Index of the tensor to get.
idx
. If the tensor with the specifiedidx
is not found, an error will return.tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_get_output_tensor(const ov_infer_request_t *infer_request, ov_tensor_t **tensor)#
Get an output tensor from the model with only one output tensor.
Note
If model has several outputs, an error will return.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
tensor – Reference to the tensor.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_infer(ov_infer_request_t *infer_request)#
Infer specified input(s) in synchronous mode.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_cancel(ov_infer_request_t *infer_request)#
Cancel inference request.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_start_async(ov_infer_request_t *infer_request)#
Start inference of specified input(s) in asynchronous mode.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_wait(ov_infer_request_t *infer_request)#
Wait for the result to become available.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_wait_for(ov_infer_request_t *infer_request, const int64_t timeout)#
Waits for the result to become available. Blocks until the specified timeout has elapsed or the result becomes available, whichever comes first.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
timeout – Maximum duration, in milliseconds, to block for.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_set_callback(ov_infer_request_t *infer_request, const ov_callback_t *callback)#
Set callback function, which will be called when inference is done.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
callback – A function to be called.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_infer_request_free(ov_infer_request_t *infer_request)#
Release the memory allocated by ov_infer_request_t.
- Parameters:
infer_request – A pointer to the ov_infer_request_t to free memory.
-
ov_infer_request_get_profiling_info(const ov_infer_request_t *infer_request, ov_profiling_info_list_t *profiling_infos)#
Query performance measures per layer to identify the most time consuming operation.
- Parameters:
infer_request – A pointer to the ov_infer_request_t.
profiling_infos – Vector of profiling information for operations in a model.
- Returns:
Status code of the operation: OK(0) for success.
-
ov_profiling_info_list_free(ov_profiling_info_list_t *profiling_infos)#
Release the memory allocated by ov_profiling_info_list_t.
- Parameters:
profiling_infos – A pointer to the ov_profiling_info_list_t to free memory.
-
struct ov_infer_request_t#
- #include <ov_infer_request.h>
type define ov_infer_request_t from ov_infer_request
-
struct ov_callback_t#
- #include <ov_infer_request.h>
Completion callback definition about the function and args.
Public Functions
- void (OPENVINO_C_API_CALLBACK *callback_func)(void *args)
The callback func.
Public Members
-
void *args#
The args of callback func.
-
struct ov_ProfilingInfo_t#
- #include <ov_infer_request.h>
Store profiling info data.
-
struct ov_profiling_info_list_t#
- #include <ov_infer_request.h>
A list of profiling info data.
Public Members
-
ov_profiling_info_t *profiling_infos#
The list of ov_profilling_info_t.
-
size_t size#
The list size.
-
ov_profiling_info_t *profiling_infos#
-
ov_infer_request_set_tensor(ov_infer_request_t *infer_request, const char *tensor_name, const ov_tensor_t *tensor)#