Neural Semantic Surface Maps
Luca Morreale 1 Noam Aigerman2,3 Vladimir G. Kim2 Niloy J. Mitra1,2
            
            1University College London 
                
            2 Adobe Research 
                
            3 University of Montreal 
        
Abstract
We present an automated technique for computing a map between two genus-zero shapes, which matches semantically corresponding regions to one another. Lack of annotated data prohibits direct inference of 3D semantic priors; instead, current state-of-the-art methods predominantly optimize geometric properties or require varying amounts of manual annotation. To overcome the lack of annotated training data, we distill semantic matches from pre-trained vision models: our method renders the pair of untextured 3D shapes from multiple viewpoints; the resulting renders are then fed into an off-the-shelf image-matching strategy that leverages a pre-trained visual model to produce feature points. This yields semantic correspondences, which are projected back to the 3D shapes, producing a raw matching that is inaccurate and inconsistent across different viewpoints. These correspondences are refined and distilled into an inter-surface map by a dedicated optimization scheme, which promotes bijectivity and continuity of the output map. We illustrate that our approach can generate semantic surface-to-surface maps, eliminating manual annotations or any 3D training data requirement. Furthermore, it proves effective in scenarios with high semantic complexity, where objects are non-isometrically related, as well as in situations where they are nearly isometric.
Method overview
Starting from a pair of upright genus-zero surfaces, we automatically distill an inter-surface map from a set of fuzzy matches. First, we align the input shapes, then extract a set of fuzzy matches through DinoV2 semantic visual features. We use these features to independently cut the two meshes and then optimize a (seamless) map between them.
        Comparisons
Here we texture only the visible region of the source model, leaving the rest white, and map it to the target shape with different techniques.
            | FAUST | SHREC07 | SHREC19 | |||||||
|---|---|---|---|---|---|---|---|---|---|
| Inv ↓ | Bij ↓ | Acc ↓ | Inv ↓ | Bij ↓ | Acc ↓ | Inv ↓ | Bij ↓ | Acc ↓ | |
| ICP | 0.06 | 0.17 | 0.25 | 0.09 | 0.65 | 0.23 | 0.07 | 0.75 | 0.15 | 
| BIM | 0.09 | 0.03 | 0.04 | 0.49 | 0.48 | 0.23 | 0.05 | 0.82 | 0.04 | 
| Zoomout | 0.33 | 0.23 | 0.15 | 0.25 | 0.65 | 0.54 | 0.29 | 0.76 | 0.32 | 
| Smooth-shells | 0.01 | 0.00 | 0.01 | 0.03 | 0.72 | 0.26 | 0.01 | 0.83 | 0.01 | 
| Ours | 0.00 | 0.00 | 0.13 | 0.00 | 0.00 | 0.23 | 0.00 | 0.00 | 0.11 | 
Bibtex
            @article{morreale2024neural,
                title={Neural Semantic Surface Maps},
                author={Morreale, Luca and Aigerman, Noam and Kim, Vladimir G. and Mitra, Niloy J.},
                booktitle={Computer Graphics Forum},
                volume={43},
                number={2},
                year={2024},
                organization={Wiley Online Library}
            }
        
      
      
