TY - GEN
T1 - DepthVoting
T2 - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024
AU - Zhu, Yunhui
AU - Chen, Jiajing
AU - Velipasalar, Senem
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Despite the significant progress in few-shot 2D image classification, few-shot 3D point cloud classification remains relatively under-explored, particularly in addressing the challenges posed by missing points in 3D point clouds. Most existing methods for few-shot 3D point cloud classification are point-based, and thus, highly sensitive to missing points. Despite recent attempts, such as ViewNet, which introduce projection-based backbones to increase robustness against missing points, the reliance on max pooling, to extract information from multiple images simultaneously, makes them prone to information loss. To address these limitations, we introduce DepthVoting, a novel projection-based approach, for few-shot 3D point cloud classification. Instead of extracting features from multiple projection images simultaneously, DepthVoting captures features from pairs of projection images (obtained from opposite view angles) separately, enhancing the extraction of more comprehensive information. These features are sent to multiple few-shot heads, which share parameters. To further refine predictions, DepthVoting incorporates a voting mechanism, allowing contribution and incorporating information from different pairs. We conduct extensive experiments on three datasets, namely ModelNet40, ModelNet40-C, and ScanObjectNN, along with cross-validation. Our proposed method consistently outperforms the state-of-the-art baselines on all datasets in terms of average accuracy with even higher margins on the challenging ScanObjectNN dataset.
AB - Despite the significant progress in few-shot 2D image classification, few-shot 3D point cloud classification remains relatively under-explored, particularly in addressing the challenges posed by missing points in 3D point clouds. Most existing methods for few-shot 3D point cloud classification are point-based, and thus, highly sensitive to missing points. Despite recent attempts, such as ViewNet, which introduce projection-based backbones to increase robustness against missing points, the reliance on max pooling, to extract information from multiple images simultaneously, makes them prone to information loss. To address these limitations, we introduce DepthVoting, a novel projection-based approach, for few-shot 3D point cloud classification. Instead of extracting features from multiple projection images simultaneously, DepthVoting captures features from pairs of projection images (obtained from opposite view angles) separately, enhancing the extraction of more comprehensive information. These features are sent to multiple few-shot heads, which share parameters. To further refine predictions, DepthVoting incorporates a voting mechanism, allowing contribution and incorporating information from different pairs. We conduct extensive experiments on three datasets, namely ModelNet40, ModelNet40-C, and ScanObjectNN, along with cross-validation. Our proposed method consistently outperforms the state-of-the-art baselines on all datasets in terms of average accuracy with even higher margins on the challenging ScanObjectNN dataset.
UR - http://www.scopus.com/inward/record.url?scp=85206469163&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85206469163&partnerID=8YFLogxK
U2 - 10.1109/CVPRW63382.2024.00074
DO - 10.1109/CVPRW63382.2024.00074
M3 - Conference contribution
AN - SCOPUS:85206469163
T3 - IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
SP - 699
EP - 707
BT - Proceedings - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024
PB - IEEE Computer Society
Y2 - 16 June 2024 through 22 June 2024
ER -