Configuration Guide for the Intel® Distribution of OpenVINO™ toolkit 2019R1 and the Intel® Vision Accelerator Design with an Intel® Arria® 10 FPGA (IEI’s Mustang-F100-A10) on Linux*¶
NOTES:
For a first-time installation, use all steps.
Use step 1 only after receiving a new FPGA card.
Repeat steps 2-4 when installing a new version of the Intel® Distribution of OpenVINO™ toolkit.
Use steps 3-4 when a Neural Network topology used by an Intel® Distribution of OpenVINO™ toolkit application changes.
1. Configure and Set Up the Intel® Vision Accelerator Design with an Intel® Arria® 10 FPGA¶
For the 2019R1.x releases, the Intel® Distribution of OpenVINO™ toolkit introduced a new board support package (BSP) a10_1150_sg1
for the Intel® Vision Accelerator Design with an Intel® Arria® 10 FPGA, which is included in the fpga_support_files.tgz
archive below. To program the bitstreams for the Intel® Distribution of OpenVINO™ toolkit 2019R1.x, you need to program the BSP into the board using the USB blaster.
Download Intel® Quartus® Prime Programmer and Tools Standard Edition 18.1. Install the Intel® Quartus® Prime Programmer and Tools Software to the
/home/<user>/intelFPGA/18.1
directory.Download
fpga_support_files.tgz
from the Intel Registration Center to the~/Downloads
directory. The files in this.tgz
archive are required to ensure your FPGA card and the Intel® Distribution of OpenVINO™ toolkit work correctly.Go to the directory where you downloaded the
fpga_support_files.tgz
archive.Unpack the
.tgz
file:tar -xvzf fpga_support_files.tgz
A directory named
fpga_support_files
is created.Go to the
fpga_support_files
directory:cd fpga_support_files
Switch to superuser:
sudo su
Use the
setup_env.sh
script fromfpga_support_files.tgz
to set your environment variables:source /home/<user>/Downloads/fpga_support_files/setup_env.sh
Uninstall any previous BSP before installing the OpenCL BSP for the 2019R1.x BSP:
aocl uninstall /opt/altera/aocl-pro-rte/aclrte-linux64/board/<BSP_package>/
Change directory to
Downloads/fpga_support_files/
:cd /home/<user>/Downloads/fpga_support_files/
Run the FPGA dependencies script, which allows OpenCL to support Ubuntu* and recent kernels:
./install_openvino_fpga_dependencies.sh
When asked, select the appropriate hardware accelerators you plan to use so it installs the correct dependencies.
If you installed the 4.14 kernel as part of the installation script, you will need to reboot the machine and select the new kernel in the Ubuntu (grub) boot menu. You will also need to rerun
setup_env.sh
to set up your environmental variables again.Export the Intel® Quartus® Prime Programmer environment variable:
export QUARTUS_ROOTDIR=/home/<user>/intelFPGA/18.1/qprogrammer
Set up the USB Blaster:
Connect the cable between the board and the host system. Use the letter codes in the diagram below for the connection points:
Connect the B end of the cable to point B on the board.
Connect the F end of the cable to point F on the FPGA download cable.
From point F end of the cable to point F on the FPGA download cable, the connection is as shown:
Run
jtagconfig
to ensure that your Intel FPGA Download Cable driver is ready to use:jtagconfig
Your output is similar to:
1) USB-Blaster [1-6] 02E660DD 10AX115H1(.|E2|ES)/10AX115H2/..
Use
jtagconfig
to slow the clock. The message “No parameter named JtagClock” can be safely ignored.jtagconfig --setparam 1 JtagClock 6M
(OPTIONAL) Confirm the clock is set to 6M:
jtagconfig --getparam 1 JtagClock
You should see the following:
6M
Go to
/opt/altera/aocl-pro-rte/aclrte-linux64/board/a10_1150_sg1/bringup
, wheresg1_boardtest_2ddr_base.sof
is located:cd /opt/altera/aocl-pro-rte/aclrte-linux64/board/a10_1150_sg1/bringup
Program the new sof file to the board:
quartus_pgm -c 1 -m JTAG -o "p;sg1_boardtest_2ddr_base.sof"
Soft reboot:
sudo reboot
Open up a new terminal and restore sudo access and the environment variables:
sudo su source /home/<user>/Downloads/fpga_support_files/setup_env.sh
Install OpenCL™ devices. Enter Y when prompted to install:
aocl install
Reboot the machine:
reboot
Open up a new terminal and restore sudo access and the environment variables:
sudo su source /home/<user>/Downloads/fpga_support_files/setup_env.sh export QUARTUS_ROOTDIR=/home/<user>/intelFPGA/18.1/qprogrammer
Run
aocl diagnose
:aocl diagnose
Your screen displays
DIAGNOSTIC_PASSED
.Use
jtagconfig
to slow the clock. The message “No parameter named JtagClock” can be safely ignored.jtagconfig --setparam 1 JtagClock 6M
Go to
/opt/intel/openvino/bitstreams/a10_vision_design_bitstreams/
, where2019R1_PL1_FP11_ResNet_SqueezeNet_VGG.aocx
is located:cd /opt/intel/openvino/bitstreams/a10_vision_design_bitstreams/
Program the
2019R1_PL1_FP11_ResNet_SqueezeNet_VGG.aocx
file to the flash to be made permanently available even after power cycle:aocl flash acl0 2019R1_PL1_FP11_ResNet_SqueezeNet_VGG.aocx
Note
You will need the USB Blaster for this.
Hard reboot the host system including powering off.
Now Soft reboot the host system to ensure the new PCIe device is seen properly
reboot
Open up a new terminal and restore sudo access and the environment variables:
sudo su source /home/<user>/Downloads/fpga_support_files/setup_env.sh
Check if the host system recognizes the Intel® Vision Accelerator Design with Intel® Arria® 10 FPGA board. Confirm you can detect the PCIe card:
lspci | grep -i Altera
Your output is similar to:
01:00.0 Processing accelerators: Altera Corporation Device 2494 (rev 01)
Run
aocl diagnose
:aocl diagnose
You should see
DIAGNOSTIC_PASSED
before proceeding to the next steps.
2. Program a Bitstream¶
The bitstream you program should correspond to the topology you want to deploy. In this section, you program a SqueezeNet bitstream and deploy the classification sample with a SqueezeNet model that you used the Model Optimizer to convert in the steps before.
Warning
Only use bitstreams from the installed version of the Intel® Distribution of OpenVINO™ toolkit. Bitstreams from older versions of the Intel® Distribution of OpenVINO™ toolkit are incompatible with later versions of the Intel® Distribution of OpenVINO™ toolkit. For example, you cannot use the 1-0-1_A10DK_FP16_Generic
bitstream, when the Intel® Distribution of OpenVINO™ toolkit supports the 2-0-1_A10DK_FP16_Generic
bitstream.
Depending on how many bitstreams you selected, there are different folders for each FPGA card type which were downloaded in the Intel® Distribution of OpenVINO™ toolkit package:
For the Intel® Vision Accelerator Design with Intel® Arria® 10 FPGA the pre-trained bistreams are in
/opt/intel/openvino/bitstreams/a10_vision_design_bitstreams
. This example uses a SqueezeNet bitstream with low precision for the classification sample.Rerun the environment setup script:
source /home/<user>/Downloads/fpga_support_files/setup_env.sh
Change to your home directory:
cd /home/<user>
Program the bitstream for the Intel® Vision Accelerator Design with Intel® Arria® 10 FPGA:
aocl program acl0 /opt/intel/openvino/bitstreams/a10_vision_design_bitstreams/2019R1_PL1_FP11_ResNet_SqueezeNet_VGG.aocx
Steps to Flash the FPGA Card¶
Note
To avoid having to reprogram the board after a power down, a bitstream will be programmed to permanent memory on the Intel® Vision Accelerator Design with Intel® Arria® 10 FPGA. This will take about 20 minutes.
The steps can be followed in the Configure and Setup the Intel® Vision Accelerator Design with an Intel® Arria® 10 FPGA section of this guide from steps 14-18 and 28-36.
3. Setup a Neural Network Model for FPGA¶
In this section, you will create an FP16 model suitable for hardware accelerators. For more information, see the FPGA plugin section in the Inference Engine Developer Guide.
Create a directory for the FP16 SqueezeNet Model:
mkdir /home/<user>/squeezenet1.1_FP16
Go to
/home/<user>/squeezenet1.1_FP16
:cd /home/<user>/squeezenet1.1_FP16
Use the Model Optimizer to convert the FP32 SqueezeNet Caffe* model into an FP16 optimized Intermediate Representation (IR). The model files were downloaded when you ran the the Image Classification verification script while installing the Intel® Distribution of OpenVINO™ toolkit for Linux* with FPGA Support. To convert, run the Model Optimizer script with the following arguments:
python3 /opt/intel/openvino/deployment_tools/model_optimizer/mo.py --input_model /home/<user>/openvino_models/models/FP32/classification/squeezenet/1.1/caffe/squeezenet1.1.caffemodel --data_type FP16 --output_dir .
The
squeezenet1.1.labels
file contains the classesImageNet
uses. This file is included so that the inference results show text instead of classification numbers. Copysqueezenet1.1.labels
to the your optimized model location:cp /home/<user>/openvino_models/ir/FP32/classification/squeezenet/1.1/caffe/squeezenet1.1.labels .
Copy a sample image to the release directory. You will use this with your optimized model:
sudo cp /opt/intel/openvino/deployment_tools/demo/car.png ~/inference_engine_samples_build/intel64/Release
4. Run a Sample Application¶
Go to the samples directory
cd /home/<user>/inference_engine_samples_build/intel64/Release
Use an Inference Engine sample to run a sample application on the CPU:
./classification_sample_async -i car.png -m ~/openvino_models/ir/FP32/classification/squeezenet/1.1/caffe/squeezenet1.1.xml
Note the CPU throughput in Frames Per Second (FPS). This tells you how quickly the inference is done on the hardware. Now run the inference using the FPGA.
Add the
-d
option to target the FPGA:./classification_sample_async -i car.png -m ~/squeezenet1.1_FP16/squeezenet1.1.xml -d HETERO:FPGA,CPU
The throughput on FPGA is listed and may show a lower FPS. This is due to the initialization time. To account for that, the next step increases the iterations to get a better sense of the speed the FPGA can run inference at.
Use
-ni
to increase the number of iterations, This option reduces the initialization impact:./classification_sample_async -i car.png -m ~/squeezenet1.1_FP16/squeezenet1.1.xml -d HETERO:FPGA,CPU -ni 100
Congratulations, you are done with the Intel® Distribution of OpenVINO™ toolkit installation for FPGA.
Additional Resources¶
Intel® Distribution of OpenVINO™ toolkit home page: https://software.intel.com/en-us/openvino-toolkit
Intel® Distribution of OpenVINO™ toolkit documentation: https://docs.openvinotoolkit.org/
Inference Engine FPGA plugin documentation: https://docs.openvinotoolkit.org/latest/_docs_IE_DG_supported_plugins_FPGA.html