December 18, 2025
Existing approaches for watermarking AI-generated outputs often rely on post-hoc methods applied in pixel space, introducing computational overhead and potential visual artifacts. In this work, we explore latent space watermarking and its distillation into generative models. We introduce DistSeal, a unified approach for latent watermarking that works across both diffusion and autoregressive models. We demonstrate that post-hoc watermarking methods can be applied directly in the latent domain and that these latent watermarkers can be effectively distilled intothe generative pipeline (either into the generative model itself or into the latent decoder) enabling in-model watermarking. The resulting post-hoc latent method achieves robustness comparable to pixel-space baselines while offering better imperceptibility and a 20× speedup. Our experiments further reveal that distilling latent watermarkers outperforms distilling pixel-space ones, providing a solution that is both more efficient and more robust. Code and model will be made available at https://github.com/facebookresearch/distseal
Written by
Sylvestre Rebuffi
Tuan Tran
Valeriu Lacatusu
Pierre Fernandez
Tomáš Souček
Tom Sander
Hady Elsahar
Alexandre Mourachko
Publisher
arxiv
Research Topics
January 02, 2026
Yuanhao Cai, Kunpeng Li, Menglin Jia, Jialiang Wang, Junzhe Sun, Feng Liang, Weifeng Chen, Felix Xu, Chu Wang, Ali Thabet, Xiaoliang Dai, Xuan Ju, Alan Yuille, Ji Hou
January 02, 2026
December 18, 2025
Aleksandar Petrov, Pierre Fernandez, Tomáš Souček, Hady Elsahar
December 18, 2025
December 18, 2025
Tomáš Souček, Pierre Fernandez, Hady Elsahar, Sylvestre Rebuffi, Valeriu Lacatusu, Tuan Tran, Tom Sander, Alexandre Mourachko
December 18, 2025
December 16, 2025
Bowen Shi, Andros Tjandra, John Hoffman, Helin Wang, Yi-Chiao Wu, Luya Gao, Julius Richter, Matt Le, Apoorv Vyas, Sanyuan Chen, Christoph Feichtenhofer, Piotr Dollar, Wei-Ning Hsu, Ann Lee
December 16, 2025

Our approach
Latest news
Foundational models