Title | ReINView: Re-interpreting Views for Multi-view 3D Object Recognition |
Authors | Xu, Ruchang Ma, Wei Mi, Qing Zha, Hongbin |
Affiliation | Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China Peking Univ, Sch Elect Engn & Comp Sci, Key Lab Machine Percept MOE, Beijing 100871, Peoples R China |
Issue Date | 2022 |
Publisher | 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS) |
Abstract | Multi-view-based 3D object recognition is important in robot-environment interaction. However, recent methods simply extract features from each view via convolutional neural networks (CNNs) and then fuse these features together to make predictions. These methods ignore the inherent ambiguities of each view caused due to 3D-2D projection. To address this problem, we propose a novel deep framework for multi-view-based 3D object recognition. Instead of fusing the multi-view features directly, we design a re-interpretation module (ReINView) to eliminate the ambiguities at each view. To achieve this, ReINView re-interprets view features patch by patch by using their context from nearby views, considering that local patches are generally co-visible at nearby viewpoints. Since contour shapes are essential for 3D object recognition as well, ReINView further performs view-level re-interpretation, in which we use all the views as context sources since the target contours to be re-interpreted are globally observable. The re-interpreted multi-view features can better reflect the 3D global and local structures of the object. Experiments on both ModelNet40 and ModelNet10 show that the proposed model outperforms state-of-the-art methods in 3D object recognition. |
URI | http://hdl.handle.net/20.500.11897/672707 |
ISBN | 978-1-6654-7927-1 |
ISSN | 2153-0858 |
DOI | 10.1109/IROS47612.2022.9981777 |
Indexed | CPCI-S(ISTP) |
Appears in Collections: | 信息科学技术学院 机器感知与智能教育部重点实验室 |