DSPoint: Dual-scale Point Cloud Recognition with High-frequency Fusion

19 Nov 2021  ·  Renrui Zhang, Ziyao Zeng, Ziyu Guo, Xinben Gao, Kexue Fu, Jianbo Shi ·

Point cloud processing is a challenging task due to its sparsity and irregularity. Prior works introduce delicate designs on either local feature aggregator or global geometric architecture, but few combine both advantages. We propose Dual-Scale Point Cloud Recognition with High-frequency Fusion (DSPoint) to extract local-global features by concurrently operating on voxels and points. We reverse the conventional design of applying convolution on voxels and attention to points. Specifically, we disentangle point features through channel dimension for dual-scale processing: one by point-wise convolution for fine-grained geometry parsing, the other by voxel-wise global attention for long-range structural exploration. We design a co-attention fusion module for feature alignment to blend local-global modalities, which conducts inter-scale cross-modality interaction by communicating high-frequency coordinates information. Experiments and ablations on widely-adopted ModelNet40, ShapeNet, and S3DIS demonstrate the state-of-the-art performance of our DSPoint.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
3D Point Cloud Classification ModelNet40 DSPoint Overall Accuracy 93.5 # 48
Number of params 1.16M # 87
Semantic Segmentation S3DIS DSPoint Mean IoU 63.3 # 38
mAcc 70.9 # 25
Number of params N/A # 1
3D Part Segmentation ShapeNet-Part DSPoint Class Average IoU 83.9 # 16
Instance Average IoU 85.8 # 36

Methods