# Optimize Semantic Segmentation Model¶

## Summary¶

INT8 Calibration is a universal method for accelerating deep learning models. Calibration is a process of converting a Deep Learning model weights to a lower 8-bit precision such that it needs less computation.

In this tutorial, you will learn how to optimize your model using INT8 Calibration, examine how much quicker the model has become, and check the difference between original and optimized model accuracy.

Model

Format

Source

Dataset

deeplabv3

Semantic Segmentation

TensorFlow*

Open Model Zoo

Pascal Visual Object Classes (Pascal VOC)

Tip

You can learn how to import a model and a dataset in the DL Workbench Get Started Guide.

Note

For deeplabv3 model, it is highly recommended to use Pascal VOC dataset to get the accurate results.

## Optimize Model Using INT8 Calibration¶

To convert the model to INT8, go to Perform tab on the Project page and open Optimize subtab. Check INT8 and click Optimize.

It takes you to the Optimize INT8 page. Select the imported dataset and perform INT8 Calibration with Default optimization method and Performance Preset calibration scheme first as it provides maximum performance speedup.

After optimization, you will be redirected to a new Project page for optimized deeplabv3 model.

To ensure that the optimized model performance is sufficiently accelerated and its predictions can be trusted, evaluate the key characteristics: performance and accuracy.

## Compare Optimized and Parent Model Performance¶

Go back to the model page and check the performance of the imported and optimized models. Compare the throughput numbers and click Compare Projects button to see more details. Learn more about projects comparison on the Compare Performance page.

Note

Throughput is the number of images processed in a given amount of time. It is measured in frames per second (FPS). Higher throughput value means better performance.

You can observe that deeplabv3 model has become 2.6x times faster on CPU device after optimization.

Lowering the precision of the model using quantization leads to a loss in prediction capability. Therefore you need to assess the model prediction capability to ensure that the model has not lost a significant amount of accuracy.

## Compare Parent and Optimized Model Predictions¶

### Create Accuracy Report¶

Create an Accuracy Report that allows you to visualize and compare Optimized and Parent model predictions. Go to the Perform tab and select Create Accuracy Report :

Comparison of Optimized and Parent Model Predictions Report allows you to find out on which validation dataset images the predictions of the model have become different after optimization. Let’s compare Optimized model predictions with Parent model predictions used as optimal references.

### Interpret Report Results¶

The Report has two display options: Basic and Advanced mode. Each line of the report table in basic mode contains an Image Name and Optimized Model Average Result for all objects in the image. Advanced mode shows Class Predicted by Optimized Model.

Basic mode:

TIP: To sort the numbers from lowest to highest, click on the parameter name in the table.

Click Visualize to see the prediction difference:

In our case, the optimized deeplabv3 model recognized all object of class #6 - buses. You can see that the clustering parts for each object coincide in Optimized and Parent model predictions.

In another example, clustering parts for each object in Optimized deeplabv3 predictions are less accurate than the Parent model prediction.

## Next Step¶

After evaluating the accuracy, you can decide whether the difference between imported and optimized models predictions is critical or not:

All images were taken from ImageNet, Pascal Visual Object Classes, and Common Objects in Context datasets for demonstration purposes only.