Set Accuracy Configuration¶
To get an adequate accuracy number, you need to correctly set accuracy configuration. Accuracy parameters depend on the model task type. To set accuracy configuration, click Provide accuracy configuration button in the Create Accuracy Report table or in the settings from the INT8 calibration tab before the optimization process or accuracy report creation. Once you have specified your parameters, you are directed back to your previous page, either the Create Accuracy Report table or the INT8 tab.
Accuracy settings depend on the model usage. Choose the usage-specific instructions from the list below:
Limitations :
DL Workbench does not support multi-input models, so make sure to use a single-input model. You can choose and download one of the Intel® Open Model Zoo models directly in the tool.
Accuracy parameters of models from Intel® Open Model Zoo are already configured. You cannot change accuracy configurations for these models.
Possible problems with accuracy :
AccuracyAware optimization method that is used to measure accuracy is impossible to apply to certain model types, for example, language models. If you see an error message when trying to calibrate such a model, return to the Calibration Options page and select the Default method.
Accuracy close to zero may appear due to incorrectly configured parameters. Often users make mistakes when setting parameters such as Color space, Normalization scales, or Normalization means in the Conversion Settings before importing the model.
Classification¶
Specify Classification in the drop-down list in the Accuracy Settings :
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV* library. |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
Accuracy (default) |
The rule that is used to compare inference results of a model with reference values. Classification accuracy metric is defined as the number of correct predictions divided by the total number of predictions. The output for a specific image is considered correct if the expected class is included into top K predictions of the model. |
Percentage |
|
Top K |
[1; 100] |
The number of top predictions among which the correct class is searched for |
N/A |
Annotation Conversion Configuration¶
Annotation conversion parameters define conversion of a dataset annotation.
Parameter |
Values |
Explanation |
---|---|---|
Separate background class |
Yes No |
Select Yes if your model was trained on a dataset with background as an additional class. Usually the index of this class is 0. |
Object Detection Single-Shot multibox Detection (SSD)¶
Specify Object Detection in the drop-down list in the Accuracy Settings :
Then specify SSD in the Model Type box that opens below.
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV library. |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Post-Processing Configuration¶
Post-processing parameters define how to process prediction values and/or annotation data after inference and before metric calculation.
Parameter |
Values |
Explanation |
---|---|---|
Prediction boxes |
None ResizeBoxes ResizeBoxes-NMS |
Resize boxes or apply Non-Maximum Suppression (NMS) to make sure that detected objects are identified only once. |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
mAP |
The rule that is used to compare inference results of a model with reference values. Mean average precision (mAP) is calculated by first finding the sum of average precisions of all classes and then dividing the sum by the number of classes. |
Percentage |
|
Metric |
COCO Precision |
The rule that is used to compare inference results of a model with reference values. COCO average precision metric is calculated by averaging precisions of all classes over Intersection over Union (IoU) values in the range from 0.50 to 0.95 with the step 0.05 . For keypoints recognition and object detection tasks, the metric is computed using bounding boxes of objects. |
Percentage |
|
Overlap threshold |
[0; 1] |
COCO precision specific. Minimal value for IoU to qualify that a detected bounding box coincides with a ground truth bounding box |
N/A |
|
Integral |
Max 11 Point |
COCO precision specific. Integral type to calculate average precision |
N/A |
|
Max Detections |
Positive Integer |
mAP-specific. Maximum number of predicted results per image. If you have more predictions, results with minimum confidence are ignored. |
N/A |
Annotation Conversion Configuration¶
Annotation conversion parameters define conversion of a dataset annotation.
Parameter |
Values |
Explanation |
---|---|---|
Separate background class |
Yes No |
Select Yes if your model was trained on a dataset with background as an additional class. Usually the index of this class is 0. |
Predictions are mapped to: |
80 COCO classes 91 COCO classes |
For COCO datasets only. Specify whether your model was trained on a dataset with 80 or 91 COCO classes. |
Object Detection You Only Look Once (YOLO) V2 and YOLO Tiny V2¶
Specify Object Detection in the drop-down list in the Accuracy Settings :
Then specify YOLO V2 or YOLO Tiny V2 in the Model Type box that opens below.
Note
YOLO models of other versions, like YOLO V3 or YOLO V5, are not supported.
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV library. |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Post-Processing Configuration¶
Post-processing parameters define how to process prediction values and/or annotation data after inference and before metric calculation.
Parameter |
Values |
Explanation |
---|---|---|
Prediction boxes |
None ResizeBoxes ResizeBoxes-NMS |
Resize boxes or apply Non-Maximum Suppression (NMS) to make sure that detected objects are identified only once. |
NMS overlap |
[0; 1] |
Non-maximum suppression overlap threshold to merge detections |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
mAP |
The rule that is used to compare inference results of a model with reference values. Mean average precision (mAP) is calculated by first finding the sum of average precisions of all classes and then dividing the sum by the number of classes. |
Percentage |
|
Metric |
COCO Precision |
The rule that is used to compare inference results of a model with reference values. COCO average precision metric is calculated by averaging precisions of all classes over Intersection over Union (IoU) values in the range from 0.50 to 0.95 with the step 0.05 . For keypoints recognition and object detection tasks, the metric is computed using bounding boxes of objects. |
Percentage |
|
Overlap threshold |
[0; 1] |
COCO precision specific. Minimal value for IoU to qualify that a detected bounding box coincides with a ground truth bounding box |
N/A |
|
Integral |
Max 11 Point |
COCO precision specific. Integral type to calculate average precision |
N/A |
|
Max Detections |
Positive Integer |
mAP-specific. Maximum number of predicted results per image. If you have more predictions, results with minimum confidence are ignored. |
N/A |
Annotation Conversion Configuration¶
Annotation conversion parameters define conversion of a dataset annotation.
Parameter |
Values |
Explanation |
---|---|---|
Separate background class |
Yes No |
Select Yes if your model was trained on a dataset with background as an additional class. Usually the index of this class is 0. |
Predictions are mapped to: |
80 COCO classes 91 COCO classes |
For COCO datasets only. Specify whether your model was trained on a dataset with 80 or 91 COCO classes. |
Instance Segmentation¶
DL Workbench supports only TensorFlow* and ONNX* instance segmentation models. ONNX instance segmentation models have different output layers for masks, boxes, predictions, and confidence scores, while TensorFlow ones have a layer for masks and a layer for boxes, predictions, and confidence scores.
Example of an ONNX instance segmentation model: instance segmentation-security-0002
Example of a TensorFlow instance segmentation model: Mask R-CNN
Specify Instance Segmentation in the drop-down list in the Accuracy Settings :
Adapter Parameters¶
Adapter parameters define conversion of inference results into a metrics-friendly format.
Parameter |
Values |
Explanation |
---|---|---|
Input info layer |
|
Name of the layer with image metadata, such as height, width, and depth |
Output layer: Masks |
boxes classes raw_masks scores |
TensorFlow-specific parameter. Boxes coordinates, predictions, and confidence scores for detected objects |
Output layer: Boxes |
boxes classes raw_masks scores |
ONNX-specific parameter. Boxes coordinates for detected objects |
Output layer: Classes |
boxes classes raw_masks scores |
ONNX-specific parameter. Predictions for detected objects |
Output layer: Scores |
boxes classes raw_masks scores |
ONNX-specific parameter. Confidence score for detected objects |
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV library. |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
COCO Segmentation Precision (default) |
The rule that is used to compare inference results of a model with reference values. COCO average precision metric for keypoints recognition and object detection tasks is calculated using masks of objects. |
Percentage |
|
Threshold start |
0.5 |
Lower threshold of the intersection over union (IoU) value |
N/A |
|
Threshold step |
0.05 |
Increment in the intersection over union (IoU) value |
N/A |
|
Threshold end |
0.95 |
Upper threshold of the intersection over union (IoU) value |
N/A |
Annotation Conversion Configuration¶
Annotation conversion parameters define conversion of a dataset annotation.
Parameter |
Values |
Explanation |
---|---|---|
Separate background class |
Yes No |
Select Yes if your model was trained on a dataset with background as an additional class. Usually the index of this class is 0. |
Semantic Segmentation¶
Specify Semantic Segmentation in the drop-down list in the Accuracy Settings :
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV library. |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Post-Processing Configuration¶
Post-processing parameters define how to process prediction values and/or annotation data after inference and before metric calculation.
Parameter |
Values |
Explanation |
---|---|---|
Segmentation mask encoding |
Annotation (default) |
Transfer mask colors to class labels using the color mapping from metadata in the annotation of a dataset. |
Segmentation mask resizing |
Prediction (default) |
Resize the model output mask to initial image dimensions. |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
Mean IoU (default) |
The rule that is used to compare inference results of a model with reference values. Mean Intersection-over-Union (mean IoU) has many flavors. For semantic segmentation, it is calculated by first computing the IoU for each semantic class and then computing the average over classes. |
Percentage |
|
Argmax |
On (default) |
Argmax is applied because the model does not perform it internally. Argmaxing is required for accuracy measurements. |
N/A |
Annotation Conversion Configuration¶
Annotation conversion parameters define conversion of a dataset annotation.
Parameter |
Values |
Explanation |
---|---|---|
Separate background class |
Yes No |
Select Yes if your model was trained on a dataset with background as an additional class. Usually the index of this class is 0. |
Predictions are mapped to: |
80 COCO classes 91 COCO classes |
For COCO datasets only. Specify whether your model was trained on a dataset with 80 or 91 COCO classes. |
Image Inpainting¶
Specify Image Inpainting in the drop-down list in the Accuracy Settings :
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Two types of masks can be applied to your image to measure its accuracy: rectangle and free form. Based on a masking type, you have two choose different sets of preprocessing parameters.
The rectangle means that there is a rectangle of specified with and height applied to the middle of the image. Example of the rectangle masking:
The free-form masking means separate lines of specified lengths, widths, and vertex numbers. Example of the free-form masking:
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV library. |
Mask type |
Rectangle Free-form |
The shape of the mask cut from an original model |
Mask width |
Positive integer |
For rectangle masking. The rectangle width in pixels |
Mask height |
Positive integer |
For rectangle masking. The rectangle height in pixels |
Number of parts |
Positive integer |
For free-form masking. The number of autogenerated forms which will be cut from an original image |
Maximum brush width |
Positive integer |
For free-form masking. The width of a form line in pixels |
Maximum length |
Positive integer |
For free-form masking. The maximum length of a form edge in pixels |
Maximum vertex count |
Positive integer greater than 2 |
For free-form masking. The maximum number of the vertices of a form |
Inverse mask |
Yes No |
If your model uses inverse masking, reset it to regular masking by checking Yes . |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
SSIM |
The rule that is used to compare inference results of a model with reference values. The structural similarity index measure (SSIM) is used to assess similarity between two images. |
Percentage |
|
Metric |
PSNR |
The rule that is used to compare inference results of a model with reference values. Peak signal-to-noise ratio (PSNR) is used as a quality measurement between the original and a modified image. Higher PSNR value means better quality of a modified image. |
Decibel |
Super-Resolution¶
Specify Super-Resolution in the drop-down list in the Accuracy Settings :
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV library. |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Annotation Conversion Configuration¶
Annotation conversion parameters define conversion of a dataset annotation.
Parameter |
Values |
Explanation |
---|---|---|
Two streams |
Yes (default) No |
Specifies whether the selected model has the second input for the upscaled image. |
Style Transfer¶
Specify Style Transfer in the drop-down list in the Accuracy Settings :
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV library. |
Normalization: mean |
[0; 255] |
Optional. The values to be subtracted from the corresponding image channels. Available for input with three channels only. |
Normalization: standard deviation |
[0; 255] |
Optional. The values to divide image channels by. Available for input with three channels only. |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
SSIM |
The rule that is used to compare inference results of a model with reference values. The structural similarity index measure (SSIM) is used to assess similarity between two images. |
Percentage |
|
Metric |
PSNR |
The rule that is used to compare inference results of a model with reference values. Peak signal-to-noise ratio (PSNR) is used as a quality measurement between the original and a modified image. Higher PSNR value means better quality of a modified image. |
Decibel |
Facial Landmark Detection¶
Specify Facial Landmark Detection in the drop-down list in the Accuracy Settings :
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV* library. |
Post-Processing Configuration¶
Post-processing parameters define how to process prediction values and/or annotation data after inference and before metric calculation.
Parameter |
Values |
Explanation |
---|---|---|
Landmark Processing |
Normalize (default) |
As a rule, a model outputs landmark coordinates in the range [0,1], while the original coordinates in a dataset correspond to the image size. To avoid mismapping, the Accuracy Checker normalizes landmark coordinates in annotations by dividing the coordinates by the image size, that is x is divided by width , and y is divided by height . |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
Normed Error (default) |
The rule that is used to compare inference results of a model with reference values. Normed error measures the quality of landmark positions. |
Percentage |
Face Recognition¶
Specify Face Recognition in the drop-down list in the Accuracy Settings :
Preprocessing Configuration¶
Preprocessing configuration parameters define how to process images prior to inference with a model.
Parameter |
Values |
Explanation |
---|---|---|
Resize type |
Auto (default) |
Automatically scale images to the model input shape using the OpenCV* library. |
Metric Configuration¶
Metric parameters specify rules to test inference results against reference values.
Parameter |
Values |
Explanation |
Units of measurement |
Details |
---|---|---|---|---|
Metric |
Pairwise Subsets (default) |
The rule that is used to compare inference results of a model with reference values. To compute pairwise accuracy, a dataset is first split into N subsets and for each subset a pairwise comparison metric is found, then the average metric across all subsets is calculated. |
Percentage |
|
Subset Count |
[2;999] |
The number of subsets N depends on the number of images M. If there are subsets that have a single image, value of the whole metric might be inadequate. Make sure to select N great enough for each subset to have at least two images. In other words, N should be not greater than M/2. |
N/A |