What is SynthID: AI Watermarking, How It Works, and Where to Use It

Last update: 2th September 2025
  • SynthID inserts imperceptible watermarks into text, images, audio, and video to enable later detection.
  • In text it is integrated as a logit processor (g-function), with configurable keys, n-grams and sampling table.
  • Verification is probabilistic with a Bayesian detector and three states: marked, unmarked, or uncertain.
  • It's not a universal standard yet; there's open source code in Transformers 4.46.0+, a discovery portal, and growing adoption.

What is SynthID

The acceleration of generative AI has accelerated the creation of text, images, audio, and video on a scale that, until recently, seemed like science fiction. In parallel, the risks of misinformation and misattributions, because it's increasingly difficult to distinguish what is human-made and what is modeled. This is where watermarking comes into play as a mitigation technique: an invisible fingerprint for the user that allows a verification system to estimate whether content was generated by AI.

Google DeepMind has introduced a family of technologies under the name SynthID that allows insert watermarks and detect them directly on the AI-generated content itself. It does this in multiple modalities (images, audio, text, and video) and, in the case of text, it has also released it as open source so any developer can implement it in their flows with libraries like Hugging Face Transformers. The result seeks greater transparency without sacrificing perceptible quality.

What is SynthID and why does it matter?

SynthID is Google's proposal to imperceptibly mark the content produced or altered by AI models and subsequently be able to identify it with a detector. The idea is not new (we already live with metadata and related standards), but here it goes a step further: the brand is embedded in the content itself —in pixels, audio waves, or token choices in text—so that it persists through common transformations and doesn’t rely solely on fragile external tags.

The motivation is clear: with viral cases like the Pope's quilted coat or the alleged arrest of Donald Trump, and with increasingly realistic generators, we need reliable clues to know where what we see comes from. Google is already applying it in its ecosystem (for example, when generating images with Gemini), and its goal is for more actors to adopt it to bring us closer to a widely shared standard in the industry.

SynthID Technology

How SynthID Works in a Nutshell

The SynthID family covers several modalities. A fingerprint is integrated into the images. directly into the pixels, imperceptible to the human eye and designed to withstand frequent changes such as cropping, applying filters, compressing or modifying the format. In audio, the brand is embedded in the signal in a way inaudible and robust against noise, MP3 compression, or speed changes. In video, it is added during frame creation and is designed to be resistant to cuts, filters, refresh rate variations or compression.

For text, SynthID Text introduces the signal during generation: it uses a sampling process that subtly modulates word choice preferences with a pseudorandom function (g-function) to encode the watermark. This modification is applied to the model's output distribution without the reader perceiving any changes in style or meaning, but maintaining a statistically detectable trace for the corresponding verifier.

SynthID Text: The Text Watermark Explained

In practical terms, SynthID Text is implemented as a logit processor which is integrated into the model generation pipeline after Top-K and Top-P. The algorithm augments the logits with a pseudorandom function g that encodes watermark information throughout the text. The goal: that a detector can later assess the probability that the text was produced with the SynthID settings, without substantially compromising quality.

The method uses a scheme of "tournament sampling» or key-guided tournament sampling. Roughly speaking, it assigns pseudo-random scores to candidate tokens, which compete against each other respecting their original probabilities, and selects the winner for each step. This controlled competition introduces a statistical signature. The research published in Nature details the technique and demonstrates that the fingerprint is difficult to forge or remove without noticeably altering the text.

  John McCarthy's Achievements: The Father of Artificial Intelligence

You don't need to retrain your model to apply it: just pass it once. brand configuration to the generation method (for example, model.generate() in Transformers) to activate SynthID's logit processor. Google and Hugging Face have facilitated a production deployment since Transformers 4.46.0 and an official testing space, plus a reference implementation on GitHub useful for contributing or porting the technique and create applications with artificial intelligence to integrate it into other projects.

Key parameters and secure settings

SynthID Text configurations have two mandatory parameters. On the one hand, Keys: a list of unique, random integers used to score the vocabulary via the g function. The length of this list defines how many "layers" of marking are applied; in practice, it is recommended to handle between 20 and 30 keys to balance quality and detectability, as suggested by the technical guidelines (see Appendix C.1 of the paper for further details).

The second essential parameter is ngram_len, which determines the n-gram window to balance robustness and detectabilityHigher values ​​tend to make detection easier, but they make the mark more vulnerable to editing; a reasonable reference value is 5. For privacy and security reasons, it is crucial that every watermark setting is stored privately and securely, because if it leaks, third parties could replicate it more easily.

In addition, you can optimize performance by using a sampling table with two properties: sampling_table_size y sampling_table_seed. To ensure a stable and unbiased g-function when sampling, a size of at least ^ 2 16, taking into account the impact on memory during inference. The seed It can be any integer, and it does not impact the perceptible quality, but it must be managed with the same care as the rest of the secrets.

Another operational detail: the repeated n-grams within the tokens of recent history (controlled by context_history_size) are excluded from the mark to improve overall detectability. This adjustment avoids reinforcing trivial patterns where the signal would be less informative.

Detection and verification: the role of the Bayesian detector

SynthID verification is probabilistic. In the Hugging Face ecosystem and on GitHub, you have a Bayesian detector that classifies each content into three states: "watermarked," "unwatermarked," or "uncertain." Using two configurable thresholds, you can adjust the relationship between false positives and false negatives, as discussed in Appendix C.8 of the technical article.

There is another operational advantage: the models that share tokenizer/parser can reuse the same brand configuration and the same detector, as long as the detector's training set includes examples of all models that will share that brand. This makes it easier to standardize a common footprint within an organization or technical consortium.

Finally, decide how to expose the detector It's part of the strategy. There are three typical options: keeping it completely private (without exposing anything), offering it in semi-private mode via an API without releasing any pesos, or making it public for general download and use. The choice depends on your infrastructure capacity and your security, transparency, and support policies.

Applications by content type: images, audio, video and text

In images, SynthID embeds an imperceptible fingerprint that does not degrade visual quality and is designed to survive regular editions such as cropping, filters, or formatting changes. This approach was initially tested with Image users (Vertex AI's text-to-image model) and has been extended to more services. In addition, the solution supports methods based on Metadata more traditional (such as those that use Photoshop-type tools), so both approaches can coexist.

  Sora AI Creating videos with text

In audio, models such as lyria integrate the branding inaudibly: it will resist added noise, MP3 compression, or speed changes. On a practical level, this helps identify podcasts created from text in Notebook LM or other formats. model-generated tracks without the user perceiving any difference in hearing.

In video, Google applies it during the own generation (for example, in Veo), with the goal of ensuring the footprint isn't affected by cropping, filters, frame rate changes, or compression. The mark doesn't alter the clip's apparent quality and aims to persist beyond a reasonable editing pipeline.

In text, as we have seen, the brand is hidden in the sequence of token decisions. In large-scale internal evaluations—more than 20 million responses with and without labeling in Gemini—it has been observed that the technique does not degrade the quality, precision, creativity or speed generation. However, its statistical strength depends on the length and how much the content is subsequently edited.

Availability, open source and ecosystem

Google has made SynthID Text available to the community as open source, with a production-grade implementation in Transformers 4.46.0+ and a Space official for testing. In addition, there is a reference implementation on GitHub useful for contributing or porting the technique to other frameworks. In the product ecosystem, the watermark already appears in Gemini (images), features like Magic Editor, and in tools like ImageFX. In video, Veo integrates SynthID and can be tested today with limited access at VideoFX.

During Google I/O 2025 it was announced SynthID Detector as a public verification portal: you upload a file (text, image, audio, or video) and it tells you if the content—or part of it—contains the brand. For now, access is limited to a small list of users, but the plan is to gradually open it up to everyone. journalists, media professionals and researchers, groups for whom this verification can be key.

Beyond the Google ecosystem, significant collaborations and adoptions have been announced. For example, NVIDIA has agreed to incorporate SynthID, a step that could accelerate its arrival in more tools in the sector. In parallel, other companies are exploring their own avenues: OpenAI, for example, has launched utilities to identify whether an image comes from DALL E and is working on advanced watermarking techniques. Convergence of approaches—or interoperability—will be crucial.

Practical implementation: cost, access and where to start

Today, SynthID is primarily integrated into premium services from Google like Vertex AI and in the Hugging Face ecosystem. It is not a "free" tool in the broad sense for production, although there are open resources (code, documentation, and examples) to experiment with and evaluate. If you develop with Transformers, you can enable SynthID Text by passing its configuration to generate(); if you work with Vertex AI (Image 2/3), the mark is applied to the exits transparently.

To test the detector, in addition to the portal announced at I/O, you have implementations in Hugging Face and GitHubIf you are an organization, you will want to plan how you will expose verification (private, API-based, or public), what performance metrics you will require (false positive/negative threshold), and how you will protect the keys configuration so that they do not leak.

Status of the standard and industry adoption

Is SynthID a universal standardNot yet. Google uses it consistently in its products and has released key pieces, but it needs other players to adopt it for verification to become ubiquitous. At best, we'll see deals between big tech companies, hardware manufacturers, and social platforms. At worst, a patchwork of incompatible solutions. For now, the movement of open source and alliances (like with NVIDIA) push in the right direction.

  What is a VPN and what is it for: 10 reasons to use a VPN

According to data shared by Google, since 2023 SynthID would have added brands to thousands of millions of content (images, videos, audio, and text). This volume hints at its potential to become a de facto reference, as regulators and public organizations push for robust solutions against deepfakes and manipulation.

Real limitations and challenges

No watermark is foolproof. SynthID Text resists clipping, moderate synonyms or minor rewordings, but their confidence drops if the text is completely rewritten or translated into another language. Furthermore, in responses very factual (where freedom of choice is limited) there is less scope for introducing signal without affecting accuracy, making it difficult to detect.

In images, audio and video, the brand supports common transformations well, but extreme manipulations can degrade the signal. SynthID is not intended to prevent a highly motivated adversary from doing damage on its own, but make it difficult and, above all, work in combination with other pieces (moderation, source validation, signed metadata, etc.) to expand coverage.

Nuances also come from the academy. Experts such as Soheil Feizi They point out that achieving a reliable and imperceptible text watermark is complex due to the nature of LLMs; however, they value the fact that the community can test detectors and assess its robustness. Others, such as Scott aaronson, underline the importance of the technique already being used applying in production, beyond theory.

Quick guide for technical teams

If you're going to evaluate it, start with Transformers 4.46.0+ and the SynthID configuration: define Keys with 20–30 unique random integers and ngram_len in 5 as a starting point. Add a sampling table with sampling_table_size ≥ 2^16 and a seed integer. Activate the processor in .generate() and generates test corpora with and without marking for train/adjust the Bayesian detector.

When calibrating the detector, define thresholds based on your tolerance to false positives and negativesEvaluate stress cases (translations, heavy paraphrasing, truncation) and check how the verifier behaves. If multiple models share a tokenizer, share them in the same brand scheme and train the detector with examples from all of them to improve its generalization.

In deployment, decide the exposure model (private, API or public), encrypt and protect the brand keys and monitor detection/latency rates. Document limits and assumptions for legal and product teams, and coordinate with communications to explain how a "uncertain" outcome.

Impact on users and creators

For the end user, having a verification portal To upload files and obtain a clear reading ("marked," "unmarked," "unclear") is a leap in transparency. For journalists and verification teams, it's an additional tool to support fact-checking practices in an environment saturated with synthetic content.

For artists and creators, having a robust footprint helps defend the authorship and provenance of their works, and distinguish the human from the synthetic when that is relevant to their audience. And for platforms, it reduces the cost of moderation, offers reliable signals for publishing policies and aligns the product with emerging regulatory requirements.

SynthID is not a magic wand, but it is a solid building block that, combined with metadata standards, cryptographic signatures and clear policies can make a difference. With increasing adoption—and if interoperability is strengthened—the sector will have a more mature tool to coexist with generative AI without sacrificing trust.

What is GitHub Spark?
Related article:
GitHub Spark: What it is and how to create applications with artificial intelligence