Coral reefs are highly vulnerable to climate change, and while photogrammetry allows for millimetre-scale mapping, supervised AI methods fail without massive labeled datasets. This ongoing research presents an approach to coral scene understanding that integrates DINOv3 semantic embeddings directly into the 3D Gaussian Splatting pipeline. By supervising high-dimensional feature vectors across multi-view imagery, we have thus far achieved qualitatively view-consistent 3D semantic segmentation. This methodology leverages the foundational strengths of Vision Transformers to provide rich, zero-shot semantic representations within a reconstructed 3D volume.
Special thanks to the creators of the wildflow/sweet-corals dataset for open-sourcing the high-quality underwater photogrammetry data used in this project. The visualisations above specifically use the Tabuhan P1 dataset.
If you find this work useful for your research, please consider citing: