Use Case and High-Level Description

This is a retrained version of the Faster R-CNN object detection network trained with the Common Objects in Context (COCO) training dataset. The actual implementation is based on Detectron, with additional network weight pruning applied to sparsify convolution layers (60% of network parameters are set to zeros).

The model input is a blob that consists of a single image of 1, 800, 1280, 3 in the BGR order. The pixel values are integers in the [0, 255] range.




Mean Average Precision (mAP)






Source framework


See Average Precision metric description at COCO: Common Objects in Context. The primary challenge metric is used. Tested on the COCO validation dataset.


Image, name: image, shape: 1, 800, 1280, 3 in the format B, H, W, C, where:

  • B - batch size

  • H - image height

  • W - image width

  • C - number of channels

Expected color order is BGR.


The net outputs a blob with the shape 1, 1, 100, 7, where each row consists of [image_id, class_id, confidence, x0, y0, x1, y1] respectively:

  • image_id - image ID in the batch

  • class_id - predicted class ID in range [1, 80], mapping to class names provided in <omz_dir>/data/dataset_classes/coco_80cl_bkgr.txt file

  • confidence - [0, 1] detection score; the higher the value, the more confident the detection is

  • (x0, y0) - normalized coordinates of the top left bounding box corner, in the [0, 1] range

  • (x1, y1) - normalized coordinates of the bottom right bounding box corner, in the [0, 1] range

Demo usage

The model can be used in the following demos provided by the Open Model Zoo to show its capabilities: