Latent Diffusion Models for Image Watermarking: A Review of Recent Trends and Future Directions
Abstract
:1. Introduction
2. Backgrounds
LDM Structure Overview
- Autoencoder (encoder and decoder): The encoder is responsible for transforming the high-dimensional input image into a compact latent representation. This step significantly reduces the data’s dimensionality, allowing for more efficient processing while retaining essential features of the original content [1]. The latent space created by the encoder is crucial as it serves as the stage for embedding watermarks without affecting pixel-level details [22]. The decoder reconstructs the image from the latent representation generated by the encoder. It is trained to reverse the transformation applied by the encoder, ensuring that the reconstructed image maintains high fidelity to the original input. Watermarking embedded in the latent space can thus be carried through the entire generation process and preserved in the output image [23].
- Latent diffusion process: The diffusion process is the heart of LDMs, involving both a forward and a reverse process: (1) Forward diffusion: In this phase, Gaussian noise is progressively added to the latent representation, simulating a diffusion process that leads to increasingly noisy versions of the latent variables. The goal of this phase is to create a path through which the model learns to handle noisy data effectively [1]. (2) Reverse diffusion (Denoising): The model is trained to reverse the forward diffusion by gradually removing noise from the latent representation. The reverse process is a stepwise denoising operation that transforms the noisy latent into a clear, usable representation, which can then be decoded into an image. The iterative nature of this process provides multiple opportunities to embed watermarks within the noise removal steps, potentially making them more resilient [22].
- UNet-based architecture: The denoising process in LDMs is typically implemented using a U-Net architecture, which is well suited for capturing both local and global features through its use of skip connections [1]. These skip connections allow the network to retain fine details from earlier layers, which is essential for high-quality image reconstruction. For watermarking purposes, the U-Net provides specific layers where watermark information could be injected during the denoising process, ensuring that it becomes an inherent part of the generative flow.
- Conditioning mechanisms: (1) Cross-attention layers: Many LDMs, such as Stable Diffusion, incorporate conditioning inputs (e.g., text prompts) to guide the generative process [1]. This conditioning is facilitated by cross-attention layers that integrate the conditioning information directly into the diffusion process. These layers provide a strategic point where watermarking can be conditioned on specific inputs, adding flexibility to the watermarking strategy [24]. (2) Latent transformation blocks: These blocks are responsible for modifying latent variables at different stages of the generation process. They present another potential point for watermark embedding, where watermarks could be introduced in a manner that integrates seamlessly with the latent transformations, making them harder to remove [24].
3. Watermarking Approaches
3.1. LDM-Based Watermarking Categorization
- (1)
- Latent space embedding (Latent): This approach involves modifying the latent space information of LDM’s variational autoencoder (VAE)-based latent space to embed a watermark. By manipulating the features within the latent representation, it is possible to incorporate watermark signals in a manner that is highly integrated and resilient to attacks [34]. This method ensures that the watermark becomes intrinsic to the latent features, resulting in a robust watermarking strategy that remains invisible to human observers.
- (2)
- Noise removal steps in the diffusion process (Diffusion Process): The iterative nature of the reverse diffusion process in LDMs offers multiple opportunities for embedding watermarks incrementally. By applying watermarking progressively at each noise removal step, the embedded information can achieve redundancy, making it harder to remove during post-processing [24]. This makes the watermark more deeply ingrained in the generated content, improving its resistance to adversarial attacks.
- (3)
- Full LDM model optimization (LDM): Another approach is to fine-tune the entire LDM model, ensuring that the watermark is inherently integrated into the outputs generated by the model. By adjusting the overall generative process, the LDM itself learns to produce images that contain embedded watermark signals [35]. This model-level optimization offers a unified watermarking solution that integrates watermarking into every generated image.
- (4)
- Decoder parameter manipulation (Decoder): This technique involves modifying the parameters of the VAE decoder within the LDM to embed a watermark. By making targeted adjustments to the decoder parameters, it is possible to introduce watermark information directly during the image reconstruction process [36]. This approach takes advantage of the final stage of image generation, ensuring that the watermark is applied as the latent representation is translated into a complete image.
- (5)
- Initial noise modification (Noise): Watermarking can also be embedded by altering the distribution of the initial noise used as input in the diffusion process. By slightly changing the characteristics of the initial noise, watermark signals can be subtly incorporated into the generated image [23]. This approach is effective as the watermark is inherently present from the start of the generative process, making it less likely to be fully removed during subsequent steps.
3.2. Watermarking Methods
3.2.1. Watermark Embedding Through Latent Space
3.2.2. Watermark Embedding via Noise Removal in Diffusion Process
3.2.3. Watermark Embedding Through Full LDM Optimization
3.2.4. Watermark Embedding via Decoder Parameter Manipulation
3.2.5. Watermark Embedding Through Initial Noise Modification
4. Technical Review and Discussion
4.1. Evaluation Metrics
4.1.1. Imperceptibility Evaluation for Watermarking in Latent Diffusion Models
- Perceptual consistency metrics: In studies such as [37], the focus is on maintaining consistency in high-level semantics rather than individual pixel values. Metrics like LPIPS are often utilized. LPIPS leverages deep neural network features to determine perceptual differences, providing a more reliable measure of how similar two images appear to a human observer when traditional pixel-based differences are not available [45].
- FID: The FID score has also been adopted as a proxy for evaluating imperceptibility in generative contexts. It measures the distance between feature representations of watermarked and non-watermarked images using an Inception network, thus quantifying how closely the distribution of watermarked images matches that of non-watermarked ones. This helps in ensuring that the introduction of a watermark does not degrade the quality of generated images perceptually [29].
- Human evaluation studies: Some works also incorporate user studies as a part of their imperceptibility evaluation. Participants are asked to differentiate between watermarked and non-watermarked images in a controlled setting to determine if watermarks are noticeable. For example, in the DiffuseTrace [21], this evaluation approach was used to empirically validate the transparency of embedded watermarks across different types of generative content.
4.1.2. Evaluation Metrics for Robustness
- Bit error rate (BER): BER is used to quantify the robustness of the watermark by measuring how many bits differ between the original and extracted watermark [46]. A low BER implies that the watermark is resilient against attacks like cropping, rotation, and compression.
- Normalized correlation (NC): NC measures the similarity between the embedded and extracted watermarks after the image has undergone attacks [46]. It helps in determining the extent to which the watermark remains intact despite modifications. A high NC value indicates that the watermark is successfully retrieved and is resistant to degradation.
- Attack simulation tests: The watermark robustness is tested under several attack scenarios, including JPEG compression, Gaussian noise addition, cropping, and brightness adjustment [46]. These tests ensure that the embedded watermark can withstand transformations that are likely to occur in real-world applications. Different studies use various image attacks, including destructive, constructive, and reconstructive attacks, to evaluate the resilience of the watermark.
4.1.3. Evaluation Metrics for Capacity
- Payload Size: Payload size is the total number of bits that can be embedded within an image. For LDM watermarking, this is a critical parameter, as a higher payload means more information can be embedded without compromising the image quality. For instance, in the Gaussian Shading method [22], the watermarking capacity is quantified in bits, and the authors demonstrate successful embedding of up to 256 bits per image without visible quality degradation.
- Embedding Rate (Bits per Pixel—bpp): This metric evaluates the density of information embedded in an image, usually expressed as bits per pixel [46]. A higher embedding rate indicates that more watermark information is present per unit of image data, which can affect robustness and image quality if not properly balanced.
4.2. Datasets and Open-Source Implementations
- MS-COCO [47]: The MS-COCO dataset, which contains over 328K images, is used extensively to evaluate the effectiveness of watermarking schemes. For LDM watermarking, a subset of 500 images is often used to assess both the imperceptibility and robustness of embedded watermarks.
- StableDiffusionDB [48]: This dataset includes images generated by Stable Diffusion based on prompts provided by users. It contains around 500 samples and is used for evaluating the quality of watermarked images under generative settings. The diverse nature of the images, influenced by user-generated prompts, allows for comprehensive robustness testing.
- Flickr30k [49]: This dataset is utilized for evaluating latent watermarking schemes. With captions provided for each image, it allows researchers to explore how well the watermark remains embedded in images generated under different textual conditions. Latent Watermark has demonstrated its effectiveness using this dataset, showing higher robustness compared to earlier methods.
5. Challenges and Open Directions
5.1. Architectural Innovations and Watermark Integration
5.2. Enhancing Robustness Against Adversarial Attacks
5.3. Balancing Imperceptibility and Detectability
5.4. Scalable Watermarking Solutions for Diverse Users
5.5. Challenges in Payload Capacity and Security
5.6. Open Directions for Enhancing Traceability and Copyright Protection
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE Computer Society: Los Alamitos, CA, USA, 2022; pp. 10684–10695. [Google Scholar]
- Brundage, M.; Avin, S.; Clark, J.; Toner, H.; Eckersley, P.; Garfinkel, B.; Dafoe, A.; Scharre, P.; Zeitzoff, T.; Filar, B.; et al. The malicious use of artificial intelligence: Forecasting, prevention, and mitigation. arXiv 2018, arXiv:1802.07228. [Google Scholar]
- Bender, E.M.; Gebru, T.; McMillan-Major, A.; Shmitchell, S. On the dangers of stochastic parrots: Can language models be too big? In Proceedings of the FAccT ’21: 2021 ACM Conference on Fairness, Accountability, and Transparency, Virtual Event, 3–10 March 2021; pp. 610–623. [Google Scholar]
- Grinbaum, A.; Adomaitis, L. The Ethical Need for Watermarks in Machine-Generated Language. arXiv 2022, arXiv:2209.03118. [Google Scholar]
- Dong, S.; Zhang, R.; Liu, J. Invisible steganography via generative adversarial network. arXiv 2018, arXiv:1807.08571. [Google Scholar]
- Singh, P.; Chadha, R.S. A survey of digital watermarking techniques, applications and attacks. Int. J. Eng. Innov. Technol. (IJEIT) 2013, 2, 165–175. [Google Scholar]
- Hosny, K.M.; Magdi, A.; ElKomy, O.; Hamza, H.M. Digital image watermarking using deep learning: A survey. Comput. Sci. Rev. 2024, 53, 100662. [Google Scholar] [CrossRef]
- Li, Y.; Wang, H.; Barni, M. A survey of deep neural network watermarking techniques. Neurocomputing 2021, 461, 171–193. [Google Scholar] [CrossRef]
- Wang, Z.; Byrnes, O.; Wang, H.; Sun, R.; Ma, C.; Chen, H.; Wu, Q.; Xue, M. Data hiding with deep learning: A survey unifying digital watermarking and steganography. IEEE Trans. Comput. Soc. Syst. 2023, 10, 2985–2999. [Google Scholar] [CrossRef]
- Amrit, P.; Singh, A.K. Survey on watermarking methods in the artificial intelligence domain and beyond. Comput. Commun. 2022, 188, 52–65. [Google Scholar] [CrossRef]
- Tsui, T.K.; Zhang, X.P.; Androutsos, D. Color image watermarking using multidimensional Fourier transforms. IEEE Trans. Inf. Forensics Secur. 2008, 3, 16–28. [Google Scholar] [CrossRef]
- Fares, K.; Amine, K.; Salah, E. A robust blind color image watermarking based on Fourier transform domain. Optik 2020, 208, 164562. [Google Scholar] [CrossRef]
- Barni, M.; Bartolini, F.; Cappellini, V.; Piva, A. A DCT-domain system for robust image watermarking. Signal Process. 1998, 66, 357–372. [Google Scholar] [CrossRef]
- Bors, A.G.; Pitas, I. Image watermarking using DCT domain constraints. In Proceedings of the 3rd IEEE International Conference on Image Processing, Lausanne, Switzerland, 19 September 1996; Volume 3, pp. 231–234. [Google Scholar]
- Al-Haj, A. Combined DWT-DCT digital image watermarking. J. Comput. Sci. 2007, 3, 740–746. [Google Scholar] [CrossRef]
- Zhu, J.; Kaplan, R.; Johnson, J.; Li, F.-F. HiDDeN: Hiding Data with Deep Networks. In Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018. [Google Scholar]
- Jia, Z.; Fang, H.; Zhang, W. MBRS: Enhancing robustness of dnn-based watermarking by mini-batch of real and simulated jpeg compression. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, 20–24 October 2021; pp. 41–49. [Google Scholar]
- Zhang, C.; Benz, P.; Karjauv, A.; Sun, G.; Kweon, I.S. UDH: Universal deep hiding for steganography, watermarking, and light field messaging. In Proceedings of the NIPS’20: 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; pp. 10223–10234. [Google Scholar]
- Zhong, X.; Huang, P.C.; Mastorakis, S.; Shih, F.Y. An automated and robust image watermarking scheme based on deep neural networks. IEEE Trans. Multimed. 2020, 23, 1951–1961. [Google Scholar] [CrossRef]
- Chen, X.; Wang, W.; Bender, C.; Ding, Y.; Jia, R.; Li, B.; Song, D. REFIT: A unified watermark removal framework for deep learning systems with limited data. In Proceedings of the ASIA CCS ’21: ACM Asia Conference on Computer and Communications Security, Virtual Event, 7–11 June 2021; pp. 321–335. [Google Scholar]
- Lei, L.; Gai, K.; Yu, J.; Zhu, L. Diffusetrace: A transparent and flexible watermarking scheme for latent diffusion model. arXiv 2024, arXiv:2405.02696. [Google Scholar]
- Yang, Z.; Zeng, K.; Chen, K.; Fang, H.; Zhang, W.; Yu, N. Gaussian Shading: Provable Performance-Lossless Image Watermarking for Diffusion Models. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 12162–12171. [Google Scholar]
- Wen, Y.; Kirchenbauer, J.; Geiping, J.; Goldstein, T. Tree-rings watermarks: Invisible fingerprints for diffusion images. In Proceedings of the NIPS’23: 37th International Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
- Min, R.; Li, S.; Chen, H.; Cheng, M. A watermark-conditioned diffusion model for ip protection. arXiv 2024, arXiv:2403.10893. [Google Scholar]
- Ci, H.; Song, Y.; Yang, P.; Xie, J.; Shou, M.Z. WMAdapter: Adding WaterMark Control to Latent Diffusion Models. arXiv 2024, arXiv:2406.08337. [Google Scholar]
- Ma, Z.; Jia, G.; Qi, B.; Zhou, B. Safe-SD: Safe and Traceable Stable Diffusion with Text Prompt Trigger for Invisible Generative Watermarking. arXiv 2024, arXiv:2407.13188. [Google Scholar]
- Zhao, Y.; Pang, T.; Du, C.; Yang, X.; Cheung, N.M.; Lin, M. A recipe for watermarking diffusion models. arXiv 2023, arXiv:2303.10137. [Google Scholar]
- Fernandez, P.; Couairon, G.; Jégou, H.; Douze, M.; Furon, T. The stable signature: Rooting watermarks in latent diffusion models. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 22466–22477. [Google Scholar]
- Varlamov, A.; Diatlova, D.; Spirin, E. METR: Image Watermarking with Large Number of Unique Messages. arXiv 2024, arXiv:2408.08340. [Google Scholar]
- Tan, Y.; Peng, Y.; Fang, H.; Chen, B.; Xia, S.T. WaterDiff: Perceptual Image Watermarks Via Diffusion Model. In Proceedings of the ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 3250–3254. [Google Scholar]
- Yuan, Z.; Li, L.; Wang, Z.; Zhang, X. Watermarking for Stable Diffusion Models. IEEE Internet Things J. 2024, 11, 35238–35249. [Google Scholar] [CrossRef]
- Kim, C.; Min, K.; Patel, M.; Cheng, S.; Yang, Y. WOUAF: Weight modulation for user attribution and fingerprinting in text-to-image diffusion models. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 8974–8983. [Google Scholar]
- Wu, H.; Liu, G.; Yao, Y.; Zhang, X. Watermarking neural networks with watermarked images. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 2591–2601. [Google Scholar] [CrossRef]
- Zhang, G.; Wang, L.; Su, Y.; Liu, A.A. A Training-Free Plug-and-Play Watermark Framework for Stable Diffusion. arXiv 2024, arXiv:2404.05607. [Google Scholar]
- Liu, Y.; Li, Z.; Backes, M.; Shen, Y.; Zhang, Y. Watermarking diffusion model. arXiv 2023, arXiv:2305.12502. [Google Scholar]
- Hu, Y.; Jiang, Z.; Guo, M.; Gong, N. Stable Signature is Unstable: Removing Image Watermark from Diffusion Models. arXiv 2024, arXiv:2405.07145. [Google Scholar]
- Meng, Z.; Peng, B.; Dong, J. Latent Watermark: Inject and Detect Watermarks in Latent Diffusion Space. arXiv 2024, arXiv:2404.00230. [Google Scholar]
- Liu, G.H.; Chen, T.; Theodorou, E.; Tao, M. Mirror diffusion models for constrained and watermarked generation. In Proceedings of the NIPS’23: 37th International Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
- Xiong, C.; Qin, C.; Feng, G.; Zhang, X. Flexible and secure watermarking for latent diffusion model. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 1668–1676. [Google Scholar]
- Desu, A.; He, X.; Xu, Q.; Lu, W. Generative Models are Self-Watermarked: Declaring Model Authentication through Re-Generation. arXiv 2024, arXiv:2402.16889. [Google Scholar]
- Rezaei, A.; Akbari, M.; Alvar, S.R.; Fatemi, A.; Zhang, Y. LaWa: Using latent space for in-generation image watermarking. arXiv 2024, arXiv:2408.05868. [Google Scholar]
- Zhang, L.; Liu, X.; Martin, A.V.; Bearfield, C.X.; Brun, Y.; Guan, H. Robust Image Watermarking using Stable Diffusion. arXiv 2024, arXiv:2401.04247. [Google Scholar]
- Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
- Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
- Hu, K.; Wang, M.; Ma, X.; Chen, J.; Wang, X.; Wang, X. Learning-based image steganography and watermarking: A survey. Expert Syst. Appl. 2024, 249, 123715. [Google Scholar] [CrossRef]
- Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common objects in context. In Computer Vision—ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
- Wang, Z.J.; Montoya, E.; Munechika, D.; Yang, H.; Hoover, B.; Chau, D.H. Diffusiondb: A large-scale prompt gallery dataset for text-to-image generative models. arXiv 2022, arXiv:2210.14896. [Google Scholar]
- Plummer, B.A.; Wang, L.; Cervantes, C.M.; Caicedo, J.C.; Hockenmaier, J.; Lazebnik, S. Flickr30k entities: Collecting region-to-phrase correspondences for richer image-to-sentence models. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 2641–2649. [Google Scholar]
- Li, X.; Shen, Q.; Kawaguchi, K. VA3: Virtually Assured Amplification Attack on Probabilistic Copyright Protection for Text-to-Image Generative Models. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 12363–12373. [Google Scholar]
- Ye, X.; Huang, H.; An, J.; Wang, Y. Duaw: Data-free universal adversarial watermark against stable diffusion customization. arXiv 2023, arXiv:2308.09889. [Google Scholar]
- Ricker, J.; Lukovnikov, D.; Fischer, A. AEROBLADE: Training-Free Detection of Latent Diffusion Images Using Autoencoder Reconstruction Error. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 9130–9140. [Google Scholar]
- Asnani, V.; Collomosse, J.; Bui, T.; Liu, X.; Agarwal, S. ProMark: Proactive Diffusion Watermarking for Causal Attribution. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 10802–10811. [Google Scholar]
- Wang, Z.; Sehwag, V.; Chen, C.; Lyu, L.; Metaxas, D.N.; Ma, S. How to Trace Latent Generative Model Generated Images without Artificial Watermark? arXiv 2024, arXiv:2405.13360. [Google Scholar]
- Zhu, P.; Takahashi, T.; Kataoka, H. Watermark-embedded Adversarial Examples for Copyright Protection against Diffusion Models. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 24420–24430. [Google Scholar]
- Nie, G.; Kim, C.; Yang, Y.; Ren, Y. Attributing image generative models using latent fingerprints. In Proceedings of the 40th International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 26150–26165. [Google Scholar]
- Megías, D.; Kuribayashi, M.; Qureshi, A. Survey on decentralized fingerprinting solutions: Copyright protection through piracy tracing. Computers 2020, 9, 26. [Google Scholar] [CrossRef]
Reference | Embedding Type | Requires Training | Controllable | Task | Fine-Tuning |
---|---|---|---|---|---|
[34] | Latent | O (Adaptor) | Adaptor | Identification | × |
[25] | Latent | O (Adaptor) | Adaptor | Identification | × |
[37] | Latent | O (Adaptor) | Adaptor | Identification | × |
[38] | Diffusion Process | X | Controllable | Verification | × |
[26] | LDM | O (Adaptor) | Prompt, Adaptor | Verification | O |
[35] | LDM | O | Prompt | Verification | O |
[31] | LDM | O | Prompt | Verification | O |
[27] | LDM | O | Prompt | Verification | O |
[39] | LDM | X | Uncontrollable | Identification | O |
[24] | LDM | X | Uncontrollable | Identification | × |
[40] | LDM | X | Uncontrollable | Verification | × |
[32] | Dec | X | Uncontrollable | Identification | O (Dec) |
[28] | Dec | X | Uncontrollable | Verification | O (Dec) |
[41] | Dec | X | Adaptor | Verification | × |
[29] | Dec, Initial Noise | X | Controllable | Identification | O (Dec) |
[29] | Initial Noise | X | Controllable | Identification | × |
[30] | Initial Noise | X | Controllable | Identification | × |
[22] | Initial Noise | X | Adaptor | Identification | × |
[23] | Initial Noise | X | Controllable | Verification | × |
[42] | Initial Noise | X | Controllable | Verification | × |
[21] | Initial Noise | X | Uncontrollable | Verification | × |
Year | Source | Model | Implementation | Links |
---|---|---|---|---|
2023 | Neurips | MDM [38] | PyTorch | https://github.com/ghliu/mdm (23 December 2024) |
Neurips | Tree [23] | PyTorch | https://github.com/YuxinWenRick/tree-ring-watermark (23 December 2024) | |
arXiv | WaterDM [27] | PyTorch | https://github.com/yunqing-me/WatermarkDM (23 December 2024) | |
ICCV | Stable [36] | PyTorch | https://github.com/facebookresearch/stable_signature (23 December 2024) | |
2024 | arXiv | METR [29] | PyTorch | https://github.com/deepvk/metr (23 December 2024) |
ECCV | WaDiff [24] | PyTorch | https://github.com/rmin2000/WaDiff (23 December 2024) | |
CVPR | Gaussian [22] | PyTorch | https://github.com/bsmhmmlf/Gaussian-Shading (23 December 2024) | |
CVPR | WOUAF [32] | PyTorch | https://github.com/kylemin/WOUAF (23 December 2024) |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hur, H.; Kang, M.; Seo, S.; Hou, J.-U. Latent Diffusion Models for Image Watermarking: A Review of Recent Trends and Future Directions. Electronics 2025, 14, 25. https://doi.org/10.3390/electronics14010025
Hur H, Kang M, Seo S, Hou J-U. Latent Diffusion Models for Image Watermarking: A Review of Recent Trends and Future Directions. Electronics. 2025; 14(1):25. https://doi.org/10.3390/electronics14010025
Chicago/Turabian StyleHur, Hongjun, Minjae Kang, Sanghyeok Seo, and Jong-Uk Hou. 2025. "Latent Diffusion Models for Image Watermarking: A Review of Recent Trends and Future Directions" Electronics 14, no. 1: 25. https://doi.org/10.3390/electronics14010025
APA StyleHur, H., Kang, M., Seo, S., & Hou, J. -U. (2025). Latent Diffusion Models for Image Watermarking: A Review of Recent Trends and Future Directions. Electronics, 14(1), 25. https://doi.org/10.3390/electronics14010025