Deep 3D Mesh Watermarking with Self-Adaptive Robustness
đĄ Research Summary
The paper introduces the first deep learningâbased framework for robust 3D mesh watermarking, addressing the longâstanding limitations of traditional handcrafted algorithms. Conventional methods fall into spatialâdomain or transformâdomain categories and must be manually tuned for each anticipated attack (e.g., cropping, remeshing, noise, smoothing). Such manual design is laborâintensive, often fails to generalize across different mesh topologies, and cannot guarantee strong robustness against all possible manipulations.
To overcome these issues, the authors propose an endâtoâend trainable network composed of three main components: (1) a watermark embedding subânetwork, (2) a set of differentiable attack layers, and (3) a watermark extraction subânetwork. The core of the architecture is a topologyâagnostic Graph Convolutional Network (GCN). Unlike standard GCNs that rely on fixed vertex ordering or edge indices, the proposed GraphConv operation uses isotropic (equal) weights for all neighboring vertices and normalizes by vertex degree. This design makes the convolution independent of mesh topology, allowing the same model to process arbitrary, nonâtemplate meshes without reâtraining.
Embedding SubâNetwork
The original mesh (V_in, F_in) and a binary watermark w_in are fed into a feature learning module consisting of five cascaded Graph Residual blocks, producing a 64âdimensional perâvertex feature map F_in. A watermark encoder compresses w_in into a latent vector z_w, which is then expanded to match the number of vertices and concatenated with the original coordinates and the learned features. The concatenated tensor passes through an aggregation module (two Graph Residual blocks plus an extra GraphConv layer) that outputs the modified vertex positions V_wm of the watermarked mesh M_wm. The âExpandingâ operation guarantees invariance to vertex reordering, a common effect of many attacks.
Attack Layers
During training, one of four simulated attacks is randomly applied to M_wm: (i) random 3âD rotation, (ii) zeroâmean Gaussian noise, (iii) Taubin smoothing, and (iv) cropping (partial mesh removal). Each attack is implemented as a differentiable layer, enabling backâpropagation through the attack. This forces the embedding and extraction networks to learn representations that survive the specific distortions, achieving what the authors term âadaptive robustness.â
Extraction SubâNetwork
The attacked mesh M_att (vertices V_att) is processed by the same feature learning module to obtain perâvertex features F_no. A global average pooling aggregates these features into a single 64âdimensional vector, which is then passed through a twoâlayer MLP to reconstruct the watermark w_ext. Global pooling ensures invariance to vertex ordering, handling the verticesâreordering attack naturally.
Loss Functions
Three losses guide training:
- Watermark loss (MSE between w_in and w_ext) enforces accurate recovery.
- Mesh distortion loss (MSE between original and watermarked vertex positions) limits overall geometric change.
- Curvature consistency loss penalizes differences in local curvature between the original and watermarked meshes, preserving visual smoothness and preventing perceptible artifacts.
Experimental Evaluation
The method is evaluated on the Stanford Bunny and ModelNet datasets under the four attack types, both individually and in mixed scenarios. Baselines include representative spatialâdomain and transformâdomain watermarking algorithms. Results show:
- Higher watermark recovery rates (5â15âŻ% absolute improvement) across all attacks, especially notable for cropping and remeshing where traditional methods often fail.
- Comparable visual quality: PSNR >âŻ38âŻdB and SSIM close to the original, thanks to the curvature loss.
- Faster processing: embedding and extraction times reduced by roughly 30âŻ% relative to optimizationâbased baselines.
- Strong transferability: a model trained on one mesh topology can be directly applied to meshes with different connectivity, maintaining >âŻ90âŻ% recovery.
Contributions
- First application of deep learning to 3D mesh watermarking, opening a new research direction.
- Introduction of a topologyâagnostic GCN that works on arbitrary meshes.
- Integration of differentiable attack layers for scenarioâspecific adaptive robustness.
- Curvatureâbased loss to guarantee imperceptibility.
- Comprehensive experiments demonstrating universal robustness, efficiency, and visual fidelity, surpassing traditional methods.
Future Work
Potential extensions include handling a broader spectrum of attacks (e.g., compression, texture alteration), scaling to higherâcapacity multiâbit watermarks, realâtime deployment within rendering pipelines, and adapting the framework to other 3D data representations such as point clouds or volumetric grids. Such advances could benefit a wide range of industriesâmedical imaging, CAD, gaming, and digital heritageâby providing robust, lowâoverhead protection of valuable 3D assets.
Comments & Academic Discussion
Loading comments...
Leave a Comment