GaussianSSC: Triplane-Guided Directional Gaussian Fields for 3D Semantic Completion
Ruiqi Xian, Jing Liang, He Yin, Xuewei Qi, Dinesh Manocha
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
We present GaussianSSC, a two-stage, grid-native and triplane-guided approach to semantic scene completion (SSC) that injects the benefits of Gaussians without replacing the voxel grid or maintaining a separate Gaussian set. We introduce Gaussian Anchoring, a sub-pixel, Gaussian-weighted image aggregation over fused FPN features that tightens voxel--image alignment and improves monocular occupancy estimation. We further convert point-like voxel features into a learned per-voxel Gaussian field and refine triplane features via a triplane-aligned Gaussian--Triplane Refinement module that combines local gathering (target-centric) and global aggregation (source-centric). This directional, anisotropic support captures surface tangency, scale, and occlusion-aware asymmetry while preserving the efficiency of triplane representations. On SemanticKITTI~behley2019semantickitti, GaussianSSC improves Stage~1 occupancy by +1.0\% Recall, +2.0\% Precision, and +1.8\% IoU over state-of-the-art baselines, and improves Stage~2 semantic prediction by +1.8\% IoU and +0.8\% mIoU.