NOTE:
EfficientNMSONNXPlugin
is deprecated since TensorRT 9.0. Its functionality has been superseded by theINMSLayer
.
This TensorRT plugin implements an efficient algorithm to perform Non Maximum Suppression for object detection networks.
This plugin is primarily intended for using with EfficientDet on TensorRT, as this network is particularly sensitive to the latencies introduced by slower NMS implementations. However, the plugin is generic enough that it will work correctly for other detections architectures, such as SSD or FasterRCNN.
The plugin has two modes of operation, depending on the given input data. The plugin will automatically detect which mode to operate as, depending on the number of inputs it receives, as follows:
-
Standard NMS Mode: Only two input tensors are given, (i) the bounding box coordinates and (ii) the corresponding classification scores for each box.
-
Fused Box Decoder Mode: Three input tensors are given, (i) the raw localization predictions for each box originating directly from the localization head of the network, (ii) the corresponding classification scores originating from the classification head of the network, and (iii) the default anchor box coordinates usually hardcoded as constant tensors in the network.
Most object detection networks work by generating raw predictions from a "localization head" which adjust the coordinates of standard non-learned anchor coordinates to produce a tighter fitting bounding box. This process is called "box decoding", and it usually involves a large number of element-wise operations to transform the anchors to final box coordinates. As this can involve exponential operations on a large number of anchors, it can be computationally expensive, so this plugin gives the option of fusing the box decoder within the NMS operation which can be done in a far more efficient manner, resulting in lower latency for the network.
Input Shape:
[batch_size, number_boxes, 4]
or[batch_size, number_boxes, number_classes, 4]
Data Type:
float32
orfloat16
The boxes input can have 3 dimensions in case a single box prediction is produced for all classes (such as in EfficientDet or SSD), or 4 dimensions when separate box predictions are generated for each class (such as in FasterRCNN), in which case number_classes
>= 1 and must match the number of classes in the scores input. The final dimension represents the four coordinates that define the bounding box prediction.
For Standard NMS mode, this tensor should contain the final box coordinates for each predicted detection. For Fused Box Decoder mode, this tensor should have the raw localization predictions. In either case, this data is given as 4
coordinates which makes up the final shape dimension.
Input Shape:
[batch_size, number_boxes, number_classes]
Data Type:
float32
orfloat16
The scores input has number_classes
elements with the predicted scores for each candidate class for each of the number_boxes
anchor boxes.
Usually, the score values will have passed through a sigmoid activation function before reaching the NMS operation. However, as an optimization, the pre-sigmoid raw scores can also be provided to the NMS plugin to reduce overall network latency. If raw scores are given, enable the score_activation
parameter so they are processed accordingly.
Input Shape:
[1, number_boxes, 4]
or[batch_size, number_boxes, 4]
Data Type:
float32
orfloat16
Only used in Fused Box Decoder mode. It is much more efficient to perform the box decoding within this plugin. In this case, the boxes input will be treated as the raw localization head box corrections, and this third input should contain the default anchor/prior box coordinates.
When used, the input must have 3 dimensions, where the first one may be either 1
in case anchors are constant for all images in a batch, or batch_size
in case each image has different anchors -- such as in the box refinement NMS of FasterRCNN's second stage.
Most input shape dimensions, namely batch_size
, number_boxes
, and number_classes
, for all inputs can be defined dynamically at runtime if the TensorRT engine is built with dynamic input shapes. However, once defined, these dimensions must match across all tensors that use them (e.g. the same number_boxes
dimension must be given for both boxes and scores, etc.)
Different object detection networks represent their box coordinate system differently. The two types supported by this plugin are:
- BoxCorners: The four coordinates represent
[x1, y1, x2, y2]
values, where each x,y pair defines the top-left and bottom-right corners of a bounding box. - BoxCenterSize: The four coordinates represent
[x, y, w, h]
values, where the x,y pair define the box center location, and the w,h pair define its width and height.
Note that for NMS purposes, horizontal and vertical coordinates are fully interchangeable. TensorFlow-trained networks, for example, often uses vertical-first coordinates such as [y1, x1, y2, x2]
, but this coordinate system will work equally well under the BoxCorner coding. Similarly, [y, x, h, w]
will be properly covered by the BoxCornerSize coding.
In Fused Box Decoder mode, the boxes and anchor tensors should both use the same coding.
The following four output tensors are generated:
-
num_detections: This is a
[batch_size, 1]
tensor of data typeint32
. The last dimension is a scalar indicating the number of valid detections per batch image. It can be less thanmax_output_boxes
. Only the topnum_detections[i]
entries innms_boxes[i]
,nms_scores[i]
andnms_classes[i]
are valid. -
detection_boxes: This is a
[batch_size, max_output_boxes, 4]
tensor of data typefloat32
orfloat16
, containing the coordinates of non-max suppressed boxes. The output coordinates will always be in BoxCorner format, regardless of the input code type. -
detection_scores: This is a
[batch_size, max_output_boxes]
tensor of data typefloat32
orfloat16
, containing the scores for the boxes. -
detection_classes: This is a
[batch_size, max_output_boxes]
tensor of data typeint32
, containing the classes for the boxes.
Type | Parameter | Description |
---|---|---|
float |
score_threshold * |
The scalar threshold for score (low scoring boxes are removed). |
float |
iou_threshold |
The scalar threshold for IOU (additional boxes that have high IOU overlap with previously selected boxes are removed). |
int |
max_output_boxes |
The maximum number of detections to output per image. |
int |
background_class |
The label ID for the background class. If there is no background class, set it to -1 . |
bool |
score_activation * |
Set to true to apply sigmoid activation to the confidence scores during NMS operation. |
bool |
class_agnostic |
Set to true to do class-independent NMS; otherwise, boxes of different classes would be considered separately during NMS. |
int |
box_coding |
Coding type used for boxes (and anchors if applicable), 0 = BoxCorner, 1 = BoxCenterSize. |
Parameters marked with a *
have a non-negligible effect on runtime latency. See the Performance Tuning section below for more details on how to set them optimally.
The EfficientNMS_ONNX_TRT
plugin's output may not always be sufficiently sized to capture all NMS-ed boxes. This is because it ignores the number of classes in the calculation of the output size (it produces an output of size (batch_size * max_output_boxes_per_class, 3)
when in general, a tensor of size (batch_size * max_output_boxes_per_class * num_classes, 3)
) would be required. This was a compromise made to keep the output size from growing uncontrollably since it lacks an attribute similar to max_output_boxes
to control the number of output boxes globally.
Due to this reason, please use TensorRT's inbuilt INMSLayer
instead of the EfficientNMS_ONNX_TRT
plugin wherever possible.
The NMS algorithm in this plugin first filters the scores below the given scoreThreshold
. This subset of scores is then sorted, and their corresponding boxes are then further filtered out by removing boxes that overlap each other with an IOU above the given iouThreshold
.
The algorithm launcher and its relevant CUDA kernels are all defined in the efficientNMSInference.cu
file.
Specifically, the NMS algorithm does the following:
-
The scores are filtered with the
score_threshold
parameter to reject any scores below the score threshold, while maintaining indexing to cross-reference these scores to their corresponding box coordinates. This is done with theEfficientNMSFilter
CUDA kernel. -
If too many elements are kept, due to a very low (or zero) score threshold, the filter operation can become a bottleneck due to the atomic operations involved. To mitigate this, a fallback kernel
EfficientNMSDenseIndex
is used instead which passes all the score elements densely packed and indexed. This method is heuristically selected only if the score threshold is less than 0.007. -
The selected scores that remain after filtering are sorted in descending order. The indexing is carefully handled to still maintain score to box relationships after sorting.
-
After sorting, the highest 4096 scores are processed by the
EfficientNMS
CUDA kernel. This algorithm uses the index data maintained throughout the previous steps to find the boxes corresponding to the remaining scores. If the fused box decoder is being used, decoding will happen until this stage, where only the top scoring boxes need to be decoded. -
The NMS kernel uses an efficient filtering algorithm that largely reduces the number of IOU overlap cross-checks between box pairs. The boxes that survive the IOU filtering finally pass through to the output results. At this stage, the sigmoid activation is applied to only the final remaining scores, if
score_activation
is enabled, thereby greatly reducing the amount of sigmoid calculations required otherwise.
The plugin implements a very efficient NMS algorithm which largely reduces the latency of this operation in comparison to other NMS plugins. However, there are certain considerations that can help to better fine tune its performance:
The algorithm is highly sensitive to the selected score_threshold
parameter. With a higher threshold, fewer elements need to be processed and so the algorithm runs much faster. Therefore, it's beneficial to always select the highest possible score threshold that fulfills the application requirements. Threshold values lower than approximately 0.01 may cause substantially higher latency.
Depending on network configuration, it is usually more efficient to provide raw scores (pre-sigmoid) to the NMS plugin scores input, and enable the score_activation
parameter. Doing so applies a sigmoid activation only to the last max_output_boxes
selected scores, instead of all the predicted scores, largely reducing the computational cost.
Some object detection networks/architectures like YOLO series need to use class-independent NMS operations. If class_agnostic
is enabled, class-independent NMS is performed; otherwise, different classes would do NMS separately.
When using networks with many anchors, such as EfficientDet or SSD, it may be more efficient to do box decoding within the NMS plugin. For this, pass the raw box predictions as the boxes input, and the default anchor coordinates as the optional third input to the plugin.
The following resources provide a deeper understanding of the NMS algorithm:
- EfficientDet
- SSD: Single Shot MultiBox Detector
- Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
- Mask R-CNN
For terms and conditions for use, reproduction, and distribution, see the TensorRT Software License Agreement documentation.
June 2023
Add deprecation note for the EfficientNMSONNXPlugin
plugin.
There are no known issues in this plugin.