HDDL Plugin¶
Introducing the HDDL Plugin¶
The Inference Engine HDDL plugin was developed for inference with neural networks on Intel Vision Accelerator Design with Intel Movidius VPUs. It is designed for use cases that require large throughput for deep learning inference, up to dozens of times more than the MYRIAD Plugin.
Configuring the HDDL Plugin¶
To configure your Intel® Vision Accelerator Design With Intel® Movidius™ on supported operating systems, refer to the Steps for Intel® Vision Accelerator Design with Intel® Movidius™ VPUs section in the installation guides for Linux or Windows.
Supported networks¶
To see the list of supported networks for the HDDL plugin, refer to the list on the MYRIAD Plugin page.
Supported Configuration Parameters¶
See VPU common configuration parameters for VPU Plugins. When specifying key values as raw strings (that is, when using the Python API), omit the KEY_
prefix.
In addition to common parameters for both VPU plugins, the HDDL plugin accepts the following options:
Parameter Name |
Parameter Values |
Default |
Description |
---|---|---|---|
KEY_PERF_COUNT |
YES/NO |
NO |
Enable performance counter option. |
KEY_VPU_HDDL_GRAPH_TAG |
string |
empty string |
Allows to execute network on specified count of devices. |
KEY_VPU_HDDL_STREAM_ID |
string |
empty string |
Allows to execute inference on a specified device. |
KEY_VPU_HDDL_DEVICE_TAG |
string |
empty string |
Allows to allocate/deallocate networks on specified devices. |
KEY_VPU_HDDL_BIND_DEVICE |
YES/NO |
NO |
Whether the network should bind to a device. Refer to ```` . |
KEY_VPU_HDDL_RUNTIME_PRIORITY |
signed int |
0 |
Specify the runtime priority of a device among all devices running the same network. Refer to ```` . |