Skip to content

Latest commit

 

History

History
36 lines (25 loc) · 2.75 KB

iou_net.md

File metadata and controls

36 lines (25 loc) · 2.75 KB

November 2019

tl;dr: Regress a separate branch to estimate the quality of object detection in terms of IoU.

Overall impression

The vanilla version of IoU-Net (with the prediction of IoU and Precise RoI Pooling) is already better than baseline, most likely due to the regularization effect of the IoU branch.

The classification confidence indicates the category of a bbox but cannot be interpreted as the localization accuracy.

It generates better results than SoftNMS (decrease the scores of overlapping ones instead of eliminating the overlapped candidate, see Review of Soft NMS on Zhihu), and can be dropped in many object detection frameworks.

KL Loss and IoU Net are similar, but are different in implmentation. KL loss directly regresses mean and var from the same head, instead of a separate head for IoU prediction in IoU Net. Also Var Voting is one forward pass, not like the IoU Net's iterative optimization.

Key ideas

  • Conventional NMS is ignorant of the localization accuracy, while the classification scores are typically used as the metric for ranking proposals.
  • localization quality is non-monotonic in iterative bounding box regression.
  • IoU-guided NMS: keep the box with the highest IoU prediction, rather than cls confidence.
  • PrRoIPooling: Use integration to make the pooling operation continuous wrt bin locations. This is an improvement over RoIAligna and RoIPool. RoIAlign only considers 2x2 point inside the bbox for calculating the feature map, but PrRoI Pooling removes this quantization altogether.
  • iterative optimization using the score estimator as a judge. Precise RoIPooling to make the backpropagation possible.
  • The regressed iou score correlates with IoU better than cls confidence

Classification confidence tends to be over-confident and is bipolar. This is similar to the results in gaussian yolov3.

Technical details

  • The IoU prediction branch is trained with random jittering online data aug.

Notes

  • The idea of using a head to regress the IoU is very much like that of FQNet, although FQNet aims to regress the 3D IoU from overlaid wireframe.
  • Iterative optimization solved the problem raised by Cascade RCNN that iterative bbox regression does not improve beyond 2 stages.
  • Also iterative optimization should be able to apply to FQNet as it has a predictor to tell the 3D IoU.