Skip to main content
Article
M3DSSD: monocular 3D single stage object detector
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
  • Shujie Luo, College of Information Science and Electronic Engineering, Zhejiang University
  • Hang Dai, Mohamed Bin Zayed University of Artificial Intelligence
  • Ling Shao, Mohamed Bin Zayed University of Artificial Intelligence
  • Yong Ding, Zhejiang University
Document Type
Conference Proceeding
Abstract

In this paper, we propose a Monocular 3D Single Stage object Detector (M3DSSD) with feature alignment and asymmetric non-local attention. Current anchor-based monocular 3D object detection methods suffer from feature mismatching. To overcome this, we propose a two-step feature alignment approach. In the first step, the shape alignment is performed to enable the receptive field of the feature map to focus on the pre-defined anchors with high confidence scores. In the second step, the center alignment is used to align the features at 2D/3D centers. Further, it is often difficult to learn global information and capture long-range relationships, which are important for the depth prediction of objects. Therefore, we propose a novel asymmetric non-local attention block with multi-scale sampling to extract depth-wise features. The proposed M3DSSD achieves significantly better performance than the monocular 3D object detection methods on the KITTI dataset, in both 3D object detection and bird's eye view tasks. The code is released at https://github.com/mumianyuxin/M3DSSD.

DOI
10.1109/CVPR46437.2021.00608
Publication Date
11-13-2021
Keywords
  • Solid modeling,
  • Computer vision,
  • Three-dimensional displays,
  • Shape,
  • Computational modeling,
  • Object detection,
  • Detectors
Comments

IR Deposit conditions: non-described

Open Access version available on CVF:

Citation Information
S. Luo, H. Dai, L. Shao and Y. Ding, "M3DSSD: monocular 3D single stage object detector," in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, 2021, pp. 6141-6150, doi: 10.1109/CVPR46437.2021.00608.