RobustPointNet++: Enhancing PointNet++ with Noise-Filtering and Learned Augmentation for Sparse LiDAR Data
Robust 3D perception in real-world applications demands models that generalize to sparse and noisy point clouds typical of LiDAR sensors. PointNet++ has achieved state-of theart performance by hierarchically capturing local geometric features, but its robustness under significant corruption remains underexplored. In this work, we integrate noise-filtering and learnable data-augmentation strategies directly into the multi scale abstraction layers of PointNet++, creating four model variants (no augmentation, jitter only, dropout only, and combined jitter+dropout). We evaluate performance on the Mod elNet40 classification benchmark under clean and synthetically corrupted conditions. Our combined-augmentation model maintains 85.68% accuracy on noisy test sets—surpassing the base network’s 82.13%—while sacrificing less than 0.66% on clean data. These results demonstrate that in-network augmentation can effectively regularize hierarchical feature learning, improving robustness to point sparsity and sensor noise. Code is available at https://github.com/ayusefi/point-cloudlearning.git.