The Foundation: What is AI Watermarking and Why We Need It

Posts

We have entered a new era of content creation. Whether it is lifelike videos and audio recordings that are indistinguishable from reality, photorealistic images generated from a simple text prompt, or engaging, well-written articles, AI-generated content is becoming increasingly ubiquitous. This technological leap, powered by sophisticated generative models, offers incredible benefits for creativity, education, and efficiency. However, this same power opens the door to serious and profound abuses, such as the rapid spread of misinformation, the targeted manipulation of public opinion, and the ability to deceive entire populations during critical events like elections. This creates a new and urgent problem: in a world where anyone can create a realistic “deepfake,” how can we know what is real?

The Provenance Problem: Misinformation and Deepfakes

Verifying the origin, or “provenance,” and authenticity of digital content has never been more important. The rise of generative models introduces a “liar’s dividend,” where bad actors can dismiss real, inconvenient footage as “just a deepfake,” while simultaneously using fake content to push a narrative. This erodes the very foundation of public trust. We need a way to distinguish between content generated by a human and content generated by a machine. We also need a way to trace a piece of content back to its source, holding creators accountable for the information they introduce into the digital ecosystem. This is not just a technical challenge; it is a societal one, and it is where the concept of AI watermarking comes into play.

Defining AI Watermarking: A New Tool for Authenticity

AI watermarking is a technique that incorporates recognizable signals, known as the watermark, directly into AI-generated content. This embedding is done in a way that makes the content traceable and protected, ideally without compromising its quality. The goal is to create a digital “fingerprint” that is intrinsically linked to the content itself. This fingerprint can then be “read” by a detection algorithm to confirm that the content was AI-generated and, in more advanced cases, to identify which specific model or even which specific user created it. This provides an essential tool for labeling and detecting AI-generated content, which is crucial for combating its misuse and building a more transparent digital environment.

Watermarking vs. Traditional Digital Watermarking

The concept of watermarking is not new. For decades, traditional digital watermarking has been used to protect copyright on images and audio. This often involved steganography, the practice of hiding a message (like a copyright notice) within a digital file. However, AI watermarking is fundamentally different. While traditional watermarking modifies a finished piece of content, AI watermarking is often integrated during the generative process itself. It is not just about hiding a logo in a file’s data. Instead, it involves subtly biasing the AI model’s output to create a statistical pattern that is imperceptible to humans but mathematically obvious to a detector. This makes the watermark part of the content’s very fabric, rather than just a superficial layer added on top.

The Core Goal: Traceability and Protection

The primary goals of AI watermarking are traceability and protection. Traceability means we can answer the question, “Where did this come from?” It allows a generative AI model developer to prove that a piece of content was or was not created by their system. This is vital for protecting their intellectual property and for taking responsibility when their models are misused. Protection means we can answer the “Is this real?” question. By providing a reliable way to identify synthetic content, watermarking helps protect the public from deception. It also protects creators, such as artists and musicians, from having their styles mimicked by AI without their consent or compensation, as a watermark can prove the origin of the synthetic media.

A Deeper Look: How AI Watermarking Differs from Metadata

One might ask why we do not simply use metadata, which is the “data about data” embedded in a file (like the “Date Created” or “Camera Model” in a photo). The reason is that metadata is extremely “brittle.” It is trivially easy for anyone to strip or alter a file’s metadata with basic software. The file itself is not changed, only the label. An AI watermark, by contrast, is part of the content itself. To remove a robust AI watermark from an image, you would have to alter the image’s pixels in such a significant way that you would likely destroy the image. A watermark is persistent; it travels with the content, even when that content is screenshotted, compressed, or re-uploaded.

The Two Faces of Watermarks: Visibility

AI watermarks can be classified according to two main factors. The first, and most obvious, is visibility. Visible watermarks are just what they sound like. They are obvious and easily recognizable elements, such as a logo or a line of text (e.g., “Created by AI Model X”) superimposed on an image or video. This approach is simple, transparent, and requires no special detector. However, it is also highly intrusive, compromises the quality of the content, and is easily defeated by simply cropping the image or editing the video to remove the visible mark. This method is good for clear labeling but bad for security and traceability.

The Invisible Signal: Imperceptible Watermarks

The real power of AI watermarking lies in imperceptible watermarks. These signals are not directly perceptible to human senses and can only be identified by specialized algorithms. They are the true digital fingerprint. In text, this might be a subtle statistical preference for certain words or sentence structures. In an image, it could be a pattern of tiny, almost invisible changes to the values or colors of specific pixels. In an audio file, it might be a slight shift in a frequency that is outside the range of human hearing. The goal of an imperceptible watermark is to provide full traceability without affecting the user’s experience or the quality of the content.

The Two Modes of Resilience: Robustness vs. Fragility

The second classification factor is resistance to manipulation. This is a critical spectrum. On one end, we have robust watermarks. These are designed to be durable. They can support and survive content modifications, whether accidental or malicious. This includes operations like compression (which happens every time you upload an image to social media), cropping, scaling, editing, or even adding filters. A robust watermark is hard to remove. On the other end, we have fragile watermarks. These watermarks are intentionally designed to be brittle. They are easily destroyed by any modification. While this sounds like a drawback, it serves a different, crucial purpose: verifying the integrity of the original, unmodified content. If the fragile watermark is still present and detectable, it serves as a guarantee that the content has not been tampered with in any way since its creation.

Why This Matters for Elections, Art, and Trust

The implications of this technology are profound. In the context of elections, the ability to reliably identify a “deepfake” audio or video of a politician could be the difference between a stable democracy and chaos. It allows social media platforms to flag and contextualize misinformation before it goes viral. In the creative fields, watermarking provides a path for generative AI artists to sign their work and for traditional artists to protect their intellectual property. Ultimately, this all comes back to the concept of trust. In our new digital reality, the old adage of “seeing is believing” is no longer true. AI watermarking is not a perfect solution, but it is one of the most promising and essential tools we have for rebuilding and maintaining trust in an online world that is increasingly filled with AI-generated content.

The Two-Step Process: A Universal Framework

To understand how AI watermarking works on a technical level, it is best to break it down into its two main steps: embedding (or encoding) and detection. This two-part framework is universal, whether the content is text, audio, images, or video. The embedding phase is where the signal is created and woven into the fabric of the content. The detection phase is where an algorithm or model is used to search for and validate the presence of that hidden signal. The specific techniques used in each phase can vary dramatically based on the type of content and the goals of the watermark, but the core process remains the same.

Phase 1: The Art of Embedding

The embedding or encoding process is the first critical step. This is where the watermark is generated and applied. This can be accomplished in various ways, ranging from simple, superficial additions to deeply integrated, generative modifications. For example, a simple method is to add subtle, low-amplitude noise patterns to an image or audio file. Another common technique is modifying the “least significant bits” (LSBs) of a file’s data. The LSB is the last bit in a string of binary data, and changing it results in such a small modification (e.g., changing a pixel’s color value from 255 to 254) that it is completely imperceptible to humans. However, these simpler methods are often not robust. The most advanced techniques involve integrating the watermark directly into the AI model’s generation process.

Watermarking Text: Subtle Linguistic Patterns

Watermarking text from a Large Language Model (LLM) is a unique challenge because text is discrete, not continuous. You cannot just change a “pixel” value. Instead, the watermark is embedded as a statistical bias. During the text generation, at certain points in a sentence, the model has to choose the next word. A watermarking algorithm will subtly bias this choice. For example, the model might be “nudged” to use a specific, pre-determined set of words (a “green list”) slightly more often than it normally would, and avoid a “red list” of words. This creates a subtle, secret linguistic pattern. To a human, the text looks perfect. But to a detection algorithm that knows the secret “green list,” this statistical anomaly is a clear and undeniable signal that the text was generated by that specific AI.

Case Study: Watermarking Large Language Models

Let’s explore this text watermarking in more detail. A common technique involves a cryptographic key. Before generating a word, the model looks at the previous few words (the context). It uses the secret key to “hash” this context, which generates a pseudo-random number. This number is then used to divide the model’s entire vocabulary into two lists: a “green list” and a “red list.” The watermarking algorithm then instructs the model to only select a word from the “green list.” This process is repeated for every new word, creating a new “green list” at each step. The result is a text that appears perfectly normal, but it has a hidden statistical property: a detector with the same secret key can check every word and see if it falls on the “green list.” A human-written text would have a 50/50 chance, but the watermarked text will have a 100% success rate, providing a mathematically provable signature.

Watermarking Images: Modifying the Pixel Fabric

For AI-generated images, watermarks are embedded by making subtle changes to the pixel values. As mentioned, the LSB method is one way, but it is fragile. A more robust method, often used by generative models, is to embed the watermark in the “latent space” of the model before the image is even created. In essence, the watermark is a part of the initial instruction, and the generative model (a GAN or diffusion model) builds the image around this hidden signal. The signal is not in one specific part of the image, but is distributed as a faint, noise-like pattern across the entire visual field. This makes it incredibly resistant to cropping; even a small piece of the image will still contain a detectable fragment of the watermark.

Watermarking Audio: Manipulating the Frequency Spectrum

Watermarking audio content involves similar principles. A common technique is “spread spectrum” watermarking, where a low-amplitude noise pattern (the watermark) is spread across a wide range of frequencies. This noise is too faint and too broad for the human ear to detect against the background of the main audio, but a detection algorithm can isolate it. Another method is to embed the watermark in frequencies that are outside the range of normal human hearing, such as in the very high or very low ranges. A more robust technique involves “echo hiding,” where a faint, slightly delayed echo of the original signal is added. The delay is so short (a few milliseconds) that the human brain perceives it as natural resonance, but a detector can identify this artificial echo pattern.

Watermarking Video: Frame-by-Frame and Encoding Adjustments

Videos present a more complex challenge, as they are a sequence of images (frames) combined with an audio stream. Watermarks can be applied to either or both. A simple method is to apply an imperceptible image watermark to each individual frame. This is effective but computationally expensive. A more advanced technique is to embed the watermark within the video’s encoding process. When a video is compressed (e.g., into an H.264 or AV1 file), the algorithm makes decisions about how to represent motion and color. A watermarking algorithm can influence these decisions, embedding a statistical pattern into the way the video itself is encoded. This makes the watermark incredibly robust, as it is part of the file’s fundamental structure.

Phase 2: The Science of Detection

The detection process is the mirror image of embedding. This is where the content is analyzed to find the hidden signal. The method of detection depends entirely on the method of embedding. For a simple LSB watermark, the detector simply reads the last bit of every byte of data to reconstruct the hidden message. For more complex, generative watermarks, the process is far more sophisticated. The detection algorithm, which often has access to a secret key, will analyze the content and look for the specific statistical anomalies it was designed to create. For a text, it checks for the “green list” pattern. For an image, it searches for the distributed noise pattern.

Machine Learning Models as Detectors

In many modern systems, the detector is not a simple algorithm; it is a machine learning model. A “classifier” model is trained to distinguish between watermarked and non-watermarked content. It is fed millions of examples of both and “learns” to identify the subtle, complex patterns that define the watermark. This is particularly useful for robust watermarks, as the detector model can be trained to recognize the watermark even after it has been damaged or distorted by compression, cropping, or other modifications. This ML-based approach is more flexible and resilient than a rigid algorithm that is only looking for a single, perfect pattern.

The Three Methods of Implementation: A Deeper Dive

Finally, the entire watermarking process can be implemented in three main ways. The first is the “generative watermark,” which is what we have mostly discussed. The watermark is embedded during the generative process itself. This is the most robust and secure method, as the watermark is fundamental to the content’s creation. The second method is the “watermark based on editing.” This is when a watermark is applied to already generated media. This is less secure, as the content exists in a “clean” state before the watermark is added, but it is a useful approach for applying watermarks to content created by models that did not have watermarking built-in. The third method is the “data-driven watermark,” where the training data of the generative model is modified, which in turn causes the model to naturally produce watermarked content. The choice of technique depends entirely on the type of content and the specific use case.

Protecting the New Gold: Intellectual Property in the AI Era

One of the most immediate and critical applications of AI watermarking is in the protection of intellectual property (IP). Generative AI models, particularly large language models and diffusion-based image models, are incredibly expensive to create. They require massive amounts of data, millions of dollars in computing power, and years of research. They are, in effect, one of the most valuable forms of intellectual property a company can own. Watermarking provides a mechanism for these companies to protect their investment. If a competitor steals a model and deploys it as their own, or if a user violates the terms of service, the watermark provides a way to prove it.

The “Radioactivity” Concept in Language Models

A fascinating study on this topic introduced the concept of “radioactivity.” This is the idea that watermarked text generated by a model is “radioactive.” If a user takes this watermarked text and uses it as training data to refine another, separate model, the “radioactivity” (the watermark) “infects” this new model. The new model will then start to produce text that also contains the original watermark’s detectable traces. This allows the developers of the original AI model to track the unauthorized reuse and “laundering” of their AI-generated content. It ensures accountability for the use of their intellectual property, as they can prove, with statistical certainty, that another model was trained on their proprietary data.

Tracking Model Theft and Unauthorized Reuse

This “radioactivity” concept is a powerful deterrent against model theft. A developer can query a competitor’s model with specific prompts and then analyze the output. If the output contains the developer’s “secret” watermark, they have a “smoking gun”—a provable, mathematical case that their model’s intellectual property was stolen and reused. This applies not just to text but to all forms of generative AI. A company’s proprietary image model, for instance, can embed a unique watermark in every image it creates. If these images are being used against the terms of service (e.g., for commercial use when only non-commercial use was permitted), the company can detect this misuse.

Case Study: Protecting Generative Art and Photography

This IP protection extends to individual creators as well. Imagine a generative artist who has spent months fine-tuning a model to produce images in their unique, signature style. They use this model to sell digital art. If someone else gains access to their model and starts selling images in the same style, the artist has little recourse. However, if the artist’s model embeds a robust, imperceptible watermark, they can prove their authorship. They can scan the web for their watermark and instantly identify all images generated by their model, allowing them to issue takedown notices or legal challenges. This same principle applies to stock photography, where a watermarked generative model can help differentiate between “official,” licensed AI stock photos and unlicensed, counterfeit ones.

The Fight Against Deepfakes: A New Arms Race

Beyond its importance for protecting intellectual property, AI watermarking plays a central and critical role in the fight against malicious deepfakes and manipulated content. This is the application with the most direct societal benefit. We are in an “arms race” between generative models that can create convincing fakes and detection models that can spot them. A watermarking system, especially one implemented at the model level by the original creator, provides an enormous advantage for the “defense.” Instead of asking a detector to find a “fake” (which is a very hard problem), we are asking it to find a “signal” (which is a much easier, more specific problem).

Verifying Authenticity in News and Media

For me, it is this dual capability for IP protection and authenticity verification that makes AI watermarking an indispensable technology. By integrating subtle and traceable markers into AI-generated content, watermarking allows us to detect manipulation and helps to maintain trust in an online world that is increasingly filled with sophisticated misinformation. News organizations, for example, could adopt a standard where all their human journalists and photographers embed a specific, fragile watermark in their work. This would serve as a seal of authenticity. If a reader sees an image or video without this watermark (or with a broken watermark), they know it has either been tampered with or did not come from that trusted source.

How Watermarking Can Restore Trust in Digital Content

This “seal of authenticity” approach can help restore trust. In a high-stakes scenario, like a political rally or a warzone, countless images and videos emerge. Bad actors will often inject “deepfake” content to sow confusion. If trusted news agencies and non-governmental organizations can adopt a secure, watermarking standard, it allows them to create a “chain of custody” for digital evidence. A fragile watermark can prove that a video has not been edited or tampered with since it was recorded. This allows journalists and human rights investigators to build a verifiable body of evidence, separating a truth from fiction.

Distinguishing Human from Machine in Creative Fields

The line between human-created and AI-generated art, music, and writing is blurring. This creates an existential problem for creative fields. Watermarking offers a path forward. It can be used not just to label AI content, but to certify human content. An AI-generated song that is watermarked as such can be appreciated for what it is. A human-made song, perhaps verified on a blockchain and linked to a fragile watermark, can be certified as a product of human creativity. This allows for new business models and ensures that human artists are not unfairly competing with a flood of unlabeled AI content. It empowers audiences to choose what they want to consume—human-made, AI-made, or a collaboration of both.

Legal and Compliance: The Role of Watermarks in Evidence

The legal field will also be profoundly impacted. As deepfake videos and audio become more common, their potential to be submitted as “evidence” in a courtroom is a serious problem. A defense lawyer could claim that a damning video of their client is just a deepfake, while a prosecutor could submit a fake video to frame a defendant. A robust watermarking standard, implemented by all major AI companies, would be a solution. A judge could require all submitted digital media to be scanned for known AI watermarks. This would allow the court to reliably distinguish between authentic digital evidence and synthetic content. This traceability is essential for maintaining justice in the digital age.

Building a Framework for Responsible AI

The development of generative AI is moving at a breathtaking pace, and the tools for governance and safety are struggling to keep up. “Responsible AI” is a framework of principles and practices designed to ensure that artificial intelligence systems are developed and deployed in ways that are safe, fair, transpart, and aligned with human values. It is about mitigating the risks of AI while maximizing its benefits. Within this framework, AI watermarking is emerging as one of the most practical and essential tools. It is not a complete solution, but it is a cornerstone of any serious AI governance strategy, providing the technical mechanism for transparency and accountability.

Watermarking as a Cornerstone of AI Governance

Given the importance of AI watermarking for verifying authenticity, I also see it as an important step in encouraging the responsible use of AI. Governance cannot exist without measurement and detection. Watermarking provides the “labeling” system that is necessary for any other policy to function. For example, a government might pass a law stating that all AI-generated content used in political advertising must be clearly labeled. This law is unenforceable without a reliable technical method to detect unlabeled AI content. Watermarking provides that method. It makes it easier to identify AI-generated content, creating a foundation for any rule, law, or policy built on top of it.

Informing the Public: The Power of a Label

One of the core principles of responsible AI is transparency. Users have a right to know if they are interacting with an AI or a human, or if the content they are consuming is real or synthetic. Watermarking is the technology that enables this transparency. A social media platform, for example, could integrate watermark detection into its upload process. When a user uploads a video, the platform scans it for known watermarks. If a watermark from a known AI-generation tool is detected, the platform can automatically apply a “AI-Generated Content” label. This simple act empowers individuals, allowing them to make informed decisions about the content they see, trust, and share, rather than being passively deceived.

Encouraging Ethical Use by Creators

Watermarking also helps keep both generative AI creators and their users informed and accountable. When a company or individual knows that the content they create with an AI tool is “stamped” with a traceable, unremovable fingerprint, their behavior changes. Generative AI creators will be more mindful of how they use AI tools, ensuring they do not mislead the public or engage in unethical practices like creating deepfakes of public figures. It creates a deterrent. The “anonymity” of creating a piece of misinformation is removed, as the content can be traced back to the model that made it. This encourages a more ethical and responsible ecosystem for AI use.

The Challenge of Industry-Wide Standardization

While the potential of AI watermarking is clear, it still faces significant challenges, notably the lack of industry-wide standards. A watermark is only useful if it can be detected. If every AI company creates its own proprietary, closed-source watermarking system, we create a digital “Tower of Babel.” A detector from one company will not be able to read the watermark from another. This makes interoperability impossible and slows wider adoption. A platform like a social media company would have to integrate dozens of different, incompatible detection algorithms, which is technically and financially unfeasible. Without a common standard, the entire promise of a universally “labeled” and “traceable” internet falls apart.

Recent Developments: A Look at Industry-Led Initiatives

Despite this challenge, I find recent developments encouraging. The major technology companies developing these models are aware of this problem and are beginning to collaborate. For example, some have introduced production-ready watermarking systems for text and images that maintain high detection accuracy with minimal impact on performance. This shows that technically, the problem is solvable at scale. Other companies have focused on video, releasing comprehensive frameworks that can insert robust signals into videos to ensure they survive transformations like compression. These individual initiatives are proving the technology’s readiness.

The Role of Open-Sourced Code in Standardization

A fascinating step in the right direction is the move by some of these major companies to publish their watermark code on public repositories. This is an interesting and crucial step towards standardization. By open-sourcing their methods, they are allowing researchers, academics, and even competitors to inspect their work, test it for vulnerabilities, and, most importantly, adopt it as a common framework. This “bottom-up” approach to standardization, where the industry converges on the best-performing open-source solution, may be faster and more effective than a “top-down” committee-based approach. It allows the community to collectively build and improve upon a shared, transparent, and auditable standard.

Interoperability: The Need for a Universal Detection Standard

The ultimate goal is interoperability. We do not need every company to use the same embedding method. We need every company to support a universal detection standard. This would be similar to how any web browser can read any website, as long as the website is written in the common language of HTML. In this future, an AI model from one company could embed its own private, proprietary watermark. However, it would also embed a second, “public” watermark that adheres to an open, industry-wide standard. This public signal would be simple, perhaps just a “yes” or “no” for “AI-Generated.” This would allow any social media platform or browser to instantly detect and label AI content, while the private watermark would still be reserved for the creator’s own IP-protection and traceability needs.

Policy and Regulation: The View from Governments

Governments and regulatory bodies are also taking notice. For anyone seeking to understand the broader challenges of AI watermarking—particularly issues related to governance, control, and the use of detection results—we are seeing major public-private discussions. For instance, the conference “What Lies Ahead for Generative AI Watermarking” organized by the European Commission brought together policymakers, researchers, and industry leaders to discuss these exact issues. This signals that watermarking is no longer just a technical nice-to-_have; it is becoming a regulatory and policy must-have. We are likely moving toward a future where regulations will mandate the watermarking of synthetic content, especially in high-stakes areas like political or financial communications.

The Debate: Is Standardization Even Possible?

While standardization is the goal, it is not without its own challenges. The field is moving so quickly that a standard set today might be obsolete in two years. Furthermore, some malicious actors, such as state-sponsored misinformation campaigns, will simply not adopt the standard. They will develop or use “rogue” AI models that do not include watermarks. Does this make the entire effort pointless? The consensus is “no.” While standardization will not stop all bad actors, it will solve the problem for the vast majority of “casually” generated misinformation. It creates a “clean” and “labeled” internet for the 99% of users who are using AI tools from responsible providers, making the “unlabeled” content from rogue actors far more conspicuous and easier to isolate and target.

The Watermark’s Dilemma: Robustness vs. Imperceptibility

While AI-powered watermarking is an extremely promising technology, it is not a silver bullet. It is crucial to understand its challenges and limitations. The most fundamental of these is the trade-off between robustness and imperceptibility. This is a constant battle. To be effective, a watermark should be “robust,” meaning it is resistant to attacks and hard to remove. It must survive the content’s journey across the internet. At the same time, the watermark must be “imperceptible,” meaning it does not impact the visual or auditory quality of the content. These two goals are in direct opposition. Improving robustness typically involves embedding the watermark more firmly or more loudly within the content, which, by its very nature, often comes at the expense of subtlety, making it more noticeable to users.

The Fragility of Subtle Watermarks

Conversely, to make a watermark truly imperceptible, it often needs to be integrated with an extremely light touch. A subtle watermark might involve changing only the least significant bit of a pixel, or a tiny statistical deviation in text. While this is great for quality, as it is completely invisible to the user, it also makes the watermark incredibly fragile. These subtle signals are often the first thing to be destroyed, even by non-malicious actions. They are vulnerable to simple compression, filtering, or even just being re-recorded. A fragile watermark is like a whisper in a hurricane; it is easily lost in the noise of the digital world, making it unreliable for traceability.

The Clumsiness of Robust Watermarks

On the other end of the spectrum, an engineer can create an incredibly robust watermark that can survive anything. They could embed a loud, complex signal across the entire frequency spectrum of an audio file. This signal would be so strong that it could survive cropping, compression, and re-recording. However, it would also be audible. It might sound like a faint hiss, a clicking, or a slight distortion. This makes the content itself lower quality and undesirable. The “perfect” watermark is one that is 100% robust and 100% imperceptible, but this is physically impossible. The real-world goal is to find the “sweet spot” in the middle that is “robust enough” for a specific use case while being “imperceptible enough” not to degrade the user’s experience.

Attack Vector 1: The Compression Gauntlet

One of the major challenges for any watermark is how its detection accuracy is affected by compression. This is perhaps the most common “attack” a watermark will face, and it is usually not even malicious. Every time you upload an image or video to a social media platform, that platform runs a compression algorithm on it. These algorithms are designed to reduce the file size by removing redundant or “less important” data. Unfortunately, a subtle, imperceptible watermark often falls into the category of “removable” or “less important” data. The compression algorithm, in its quest for efficiency, can inadvertently “clean” the content of its watermark, making the AI-generated content undetectable by the time it is viewed by others.

Attack Vector 2: Cropping, Resizing, and Scaling

Simple transformations are another major threat. Cropping an image, for example, takes this challenge even further. If a watermark is embedded as a logo in the corner of an image (even an invisible one), a simple crop will remove it entirely. This is why more advanced watermarks are distributed as a faint, noise-like pattern across the entire image. This way, even a small cropped piece of the image still contains a fragment of the watermark. However, this fragmented signal can be much harder to detect. Resizing and scaling pose a similar problem. When an image is shrunk, the algorithm discards pixels, and it may discard the very pixels that contain the watermark signal.

Attack Vector 3: The Paraphrasing Attack on Text

Text watermarks, which rely on statistical patterns, have their own unique and significant vulnerability: paraphrasing. If a user generates a block of watermarked text from an AI, they can easily “launder” it by simply rephrasing the sentences. They might use another AI model (a “paraphrasing tool”) or do it manually. This “attack” breaks the original word-choice patterns that the watermark relies on, effectively destroying the signal. A user could even copy the text, paste it into a different AI model, and ask it to “rewrite this in a different style.” The new text would contain all the same information and ideas, but it would have the statistical fingerprint of the second model, or no fingerprint at all, completely removing the traceability of the original.

Attack Vector 4: The Adversarial Attack

A more sophisticated and malicious attack is an “adversarial attack.” This is when an attacker knows that a watermark exists and actively tries to remove it. If a watermarking algorithm is public or if its principles are well-understood, an attacker can design a tool to specifically target it. This tool would be a machine learning model of its own, trained to analyze a piece of content, identify the subtle statistical patterns of the watermark, and then apply the minimum possible modifications to “erase” that pattern while preserving the content’s quality. This becomes a high-stakes, cat-and-mouse game between the watermark embedder and the adversarial attacker.

Can Watermarks Survive the “Laundry” Process?

These attacks lead to the concept of a “watermark laundry.” This is a hypothetical (or, in some cases, real) process designed to “clean” content of its watermark. This “laundry” could be a simple, multi-step process: take an AI-generated image, compress it, resize it, add a filter, and then re-upload it. This chain of transformations, which is very common online, can be enough to destroy most fragile watermarks. A more advanced “laundry” would be an adversarial tool that is specifically designed to “wash” content by rebuilding it. For example, a tool could take a watermarked video, break it down into individual frames, use an AI model to “re-paint” each frame, and then re-assemble the video, creating a new, “clean” copy that is visually identical but has no watermark.

The Latency and Cost Trade-Off

A final, practical limitation is performance. Watermarking, especially at the generative-process level, is not “free.” It adds a small amount of computational overhead to the AI model. The watermarking algorithm has to run on every single piece of content the model generates. For a text model, this might be a few extra milliseconds per word. For a video model, it could be more. This “latency” can be a problem for real-time applications, like a live-translation service or a conversational AI chatbot. Companies must weigh the need for security (watermarking) against the need for speed. This also translates to cost, as the extra computation required for robust watermarking can add up to millions of dollars in server costs for a large-scale service.

The Scalability Problem: Watermarking a Trillion-Token Model

This challenge of cost and latency is part of a broader scalability problem. The AI models being developed are massive, and they are serving billions of users. The watermarking solution must be able to scale with them. It needs to be lightweight enough not to slow down the user experience, yet robust enough to be effective. It needs to be cheap enough to run on every single one of the trillions of pieces of content being generated. And it needs to be future-proof, capable of being adapted as the generative models themselves become even larger and more complex. Finding a single technique that is robust, imperceptible, fast, cheap, and scalable is the central, unsolved challenge in the field of AI watermarking.

The Horizon of AI Watermarking

When I think about the future of AI watermarking, I see several exciting and critical advances on the horizon. The technology is still in its relative infancy, and as generative AI models become more complex, the watermarking techniques designed to trace them must evolve as well. The future will likely move toward more sophisticated integration with the models themselves, new detection methods, and a much more complex and necessary conversation about the ethics of traceability. This is not just a technical evolution; it is a societal one, as we decide how to balance the need for trust with the right to privacy.

Techniques Inspired by Cryptography: The Secret Key

One of the most interesting and powerful approaches on the horizon is the use of techniques inspired by cryptography. In most current watermarking systems, the “secret” is the method itself. In a cryptographic approach, the method is public, but its implementation is secured by a “secret key.” This means a watermark is embedded using a secret key that is known only to the model’s creator. The key difference is in detection. Without the secret key, it is computationally impossible to distinguish the watermarked content from the original, non-watermarked content. The signal is not just hidden; it is cryptographically secured.

The “Undetectable” Watermark Concept

This “secret key” approach leads to the concept of “undetectable” watermarks. This may sound counter-intuitive, but it is a revolutionary idea. The goal is to create a watermark that is undetectable to anyone except the owner of the key. This solves a major security flaw of current systems. If a public detector exists for a watermark, an attacker can use that detector to their advantage. They can modify the content, run it through the detector, and repeat the process until the detector says the content is “clean.” A cryptographic, “secret key” watermark is immune to this attack. An attacker, lacking the key, has no way of knowing if their modifications have successfully removed the signal. For more in-depth information on this, I highly recommend consulting the research on undetectable watermarks for language models.

The Ethical Tightrope: Transparency vs. Privacy

However, while these advances are exciting, they bring with them significant ethical concerns about freedom of expression, privacy, and transparency. A perfectly robust and traceable watermark is a double-edged sword. It can be used to spot misinformation, but it can also be used as a tool for surveillance. This creates a difficult ethical tightrope. While we want to stop the spread of deepfakes, do we also want to live in a world where every piece of AI-generated text or art is traceable back to its creator? This balance between transparency (knowing what is real) and privacy (the right to create anonymously) is perhaps the most difficult challenge of all.

The Human Rights Defender: A Use Case in Privacy

Consider the example provided in the original article, which is a powerful one. An image is generated by a human rights defender to document an act of abuse, perhaps using an AI tool to anonymize the faces of victims. If that image contains a hidden, robust watermark, it could be used by an oppressive regime to identify the human rights defender, placing them in grave danger. In this context, traceability is a bug, not a feature. This is not a hypothetical problem. The same technology that can be used to identify a “troll” spreading misinformation can be used to identify a “dissident” sharing a truth.

The Chilling Effect on Freedom of Expression

This leads to a broader concern about the “chilling effect” on freedom of expression. If creators know that their every AI-assisted output is being watermarked and tracked, it may stifle creativity and experimentation. People may become afraid to explore sensitive or controversial topics, even with the assistance of AI, for fear of being misidentified or having their work taken out of context. It is critically important that AI developers and policymakers address this issue. We must ensure that watermarks are designed to protect the privacy of people who create and share sensitive content, while also enabling effective attribution and traceability. This might mean creating “opt-out” systems or developing “privacy-preserving” watermarks that can prove content is synthetic without revealing the identity of the creator.

The Future of Detection: Advanced Neural Networks

The future of detection will also be more advanced. As generative models become more complex, the watermarks and the detectors will become more complex as well. The detection process will likely not be a simple algorithm but another, specialized neural network. This “detector AI” will be trained to recognize the subtle, high-dimensional statistical “fingerprints” of various generative models. This could lead to a future where a web browser or operating system has a built-in AI detection layer. As you browse the internet, this detector would be running in the background, scanning all the content you see and providing you with a real-time “trust score” for every image, video, and article.

The Policy Challenge: Who Gets to Detect?

This advanced detection capability raises another difficult policy question: who gets to detect? If the “secret key” approach becomes the standard, who gets to hold the key? Only the creating company? What if law enforcement demands the key to track a suspect? What if a court subpoenas the key to verify evidence? And if detection is handled by a few large tech platforms, does that give them too much power to act as the “arbiters of truth”? The governance of the detection infrastructure may become just as important as the governance of the AI generation models themselves.

A Symbiotic Future: Watermarking and AI Co-Evolution

The relationship between artificial intelligence systems capable of generating synthetic content and the watermarking technologies designed to identify that content represents one of the most fascinating and consequential dynamics in contemporary technology. This relationship transcends simple opposition between creation and detection, evolving instead into a complex interplay where advances on either side drive corresponding innovations on the other, creating an ongoing cycle of adaptation and counter-adaptation that promises to continue indefinitely. Understanding this dynamic proves essential not just for technologists working on these systems but for policymakers, platform operators, and society more broadly as we collectively navigate a future where synthetic and authentic content increasingly intermingle.

The trajectory of this relationship reveals patterns familiar from other domains involving adversarial dynamics, from biological evolution to cybersecurity to financial fraud detection. In each of these domains, competing forces drive continuous innovation as advances on one side create pressure for corresponding advances on the other. This adversarial innovation produces neither stable equilibrium nor ultimate victory for either side but rather an ongoing process of adaptation where the state of capability continuously shifts without ever reaching a final resolution. Recognizing that AI content generation and watermarking detection exist within this same pattern of perpetual evolution helps set appropriate expectations and guides more realistic approaches to policy and system design.

The Fundamental Dynamics of Co-Evolution

The concept of co-evolution, borrowed from biology where it describes how species evolve in response to each other, provides a useful framework for understanding the relationship between generative AI systems and watermarking technologies. Just as predators and prey evolve in response to each other’s adaptations, with improvements in predator capabilities driving evolution of better prey defenses which in turn drives further predator evolution, generative AI and watermarking detection drive each other’s advancement through competitive pressure.

This co-evolutionary dynamic operates through several mechanisms that ensure continuous innovation on both sides. When watermarking techniques improve in their ability to reliably detect AI-generated content, this success creates immediate pressure on generative systems to evolve approaches that evade detection. System developers, particularly those seeking to use AI-generated content in contexts where detection would be problematic, have strong incentives to innovate around detection capabilities. These innovations might involve subtle modifications to generation processes that preserve content quality while disrupting watermarking signals, post-processing techniques that remove or obscure watermarks, or entirely new generation architectures designed from the ground up to resist watermarking.

Conversely, when generative systems advance in their ability to evade watermarking detection, this success creates pressure on detection technologies to evolve more robust approaches. Researchers and developers working on watermarking systems study the evasion techniques employed by generative systems and develop countermeasures that restore detection capability. These countermeasures might involve more sophisticated signal analysis, multiple redundant watermarking approaches that prove difficult to simultaneously evade, or fundamentally new approaches to detection that do not rely on traditional watermarking signals.

The competitive pressure driving innovation on both sides ensures that neither capability remains static for long. Breakthroughs in generative quality or watermarking detection quickly trigger responsive innovations on the opposing side. The pace of this innovation cycle depends on factors including the intensity of incentives driving development, the resources available to researchers on both sides, the difficulty of the technical challenges involved, and the existence of fundamental limits on what is possible for either generation or detection.

The asymmetries between generation and detection also shape the co-evolutionary dynamics. In many cases, evading detection proves easier than achieving reliable detection, as attackers need only find a single successful evasion approach while defenders must guard against all possible evasion techniques. This asymmetry can create situations where detection capabilities lag behind evasion techniques despite significant defensive investment. However, other factors can favor detection, including the ability to learn from observed evasion attempts, the possibility of detection approaches that exploit fundamental properties of generation processes, and the collective nature of defensive effort where multiple parties contribute to detection research.

Increasing Sophistication on Both Sides

The co-evolutionary pressure drives increasing sophistication in both generative AI systems and watermarking technologies, producing capabilities that would not emerge without the competitive dynamic. This increasing sophistication manifests in multiple dimensions including technical complexity, robustness to adversarial manipulation, and adaptability to evolving challenges.

On the generation side, sophistication increases through multiple avenues of innovation. Model architectures evolve to produce higher quality output that more closely resembles authentic human-created content, making detection based on quality differences increasingly difficult. Generation processes incorporate techniques specifically designed to resist watermarking, such as adding controlled randomness that disrupts watermarking signals while preserving output quality, or using ensemble approaches that combine multiple models in ways that confound detection systems trained on individual models.

Advanced generation systems also incorporate adaptive capabilities that enable them to respond to detection techniques in real time. These systems might test generated content against available detectors and modify generation parameters when detection risk appears high, iteratively refining outputs until they evade detection. They might employ adversarial training approaches where detection systems are used during training to produce models inherently resistant to detection. The sophistication of these adaptive approaches continues increasing as developers apply more advanced machine learning techniques to the challenge of evasion.

On the detection side, watermarking technologies evolve corresponding sophistication through complementary innovations. Detection systems move beyond simple signal detection to employ advanced analysis that can identify subtle patterns indicating AI generation even when obvious watermarks are absent or have been manipulated. Machine learning approaches enable detection systems to learn from large collections of both authentic and AI-generated content, identifying statistical regularities that distinguish the two even as generation quality improves.

Robust watermarking techniques that resist removal or manipulation represent another avenue of increasing sophistication. These approaches embed watermarking signals deeply within content structure rather than superficially, making removal difficult without severely degrading content quality. They employ redundancy where multiple independent watermarking signals must all be removed or manipulated for evasion to succeed. They leverage fundamental properties of generation processes that cannot be easily modified without fundamentally changing model architectures.

The increasing sophistication on both sides creates a situation where both generation and detection capabilities far exceed what existed even a few years ago, yet the fundamental challenge of reliably detecting AI-generated content remains unresolved. Generation systems can produce content of remarkable quality that evades many detection approaches, while detection systems can identify AI generation with reasonable reliability in many contexts. Neither side has achieved dominant superiority that would end the competitive dynamic, ensuring that co-evolution continues.

The Adversarial Innovation Cycle

The pattern of innovation in AI watermarking exemplifies the broader phenomenon of adversarial innovation that appears across many domains involving security, authentication, or verification. This pattern, while frustrating for those seeking permanent solutions to detection challenges, represents an inevitable feature of systems where opposing parties have strong incentives to outmaneuver each other and where the technological landscape permits continuous innovation.

The adversarial cycle typically follows a recurring pattern. A new watermarking or detection technique emerges that successfully identifies AI-generated content with high reliability. This success creates a period where detection appears to have gained the upper hand, leading to optimism about achieving robust identification of synthetic content. However, this success also creates strong incentives for those seeking to evade detection to study the new technique and develop countermeasures. Researchers and developers analyze how the detection works, identify its assumptions and limitations, and design generation approaches that exploit these weaknesses.

When successful evasion techniques emerge, detection capability degrades and the advantage shifts to generation systems. This shift creates pressure on detection researchers to understand the evasion techniques and develop countermeasures that restore detection capability. The cycle then repeats with new detection approaches, subsequent evasion techniques, and further countermeasures, creating an ongoing oscillation where neither side maintains permanent advantage.

The pace of this adversarial cycle depends on multiple factors. Strong economic or strategic incentives to evade detection accelerate the development of evasion techniques, as when malicious actors seek to use AI-generated content for fraud, disinformation, or other harmful purposes. Conversely, strong incentives to maintain reliable detection, whether from platforms seeking to enforce content policies, regulators requiring identification of synthetic content, or public interest in authenticity verification, drive investment in detection research.

The technical difficulty of evasion or detection also influences cycle pace. When fundamental properties of generation processes enable robust watermarking that cannot be easily removed without severe quality degradation, detection may maintain advantage for extended periods. Conversely, when watermarking techniques rely on fragile signals easily disrupted by minor modifications, evasion may dominate. The balance of technical difficulty between evasion and detection varies across different types of content and generation approaches, creating different dynamics in different domains.

The openness or secrecy surrounding techniques also affects adversarial dynamics. When detection techniques are published openly, evasion researchers can study them directly and develop targeted countermeasures. When techniques remain proprietary or secret, evasion must proceed more slowly through trial-and-error exploration. However, security through obscurity proves fragile in the long run, as determined adversaries eventually uncover even secret techniques through reverse engineering or other means.

Why Perfect Solutions Remain Elusive

The recognition that adversarial innovation creates perpetual cycles rather than final victories raises the question of whether fundamental limitations prevent achievement of perfect solutions to watermarking and detection challenges. Understanding why such perfect solutions remain elusive helps set realistic expectations and guides more effective approaches to the challenge of identifying AI-generated content.

Several fundamental factors conspire to prevent perfect solutions. The first involves the underlying nature of the challenge, which pits capabilities that exist at roughly the same level of technological sophistication against each other. Both generation and detection employ advanced machine learning, sophisticated signal processing, and substantial computational resources. When opponents possess comparable capabilities, achieving decisive permanent advantage becomes extremely difficult.

The second factor involves the availability of multiple approaches to both generation and detection. The space of possible generation techniques is vast, encompassing different model architectures, training approaches, sampling strategies, and post-processing methods. Similarly, detection can employ diverse approaches including watermarking signal detection, statistical analysis, model fingerprinting, and behavioral pattern recognition. This multiplicity means that even when one approach to evasion or detection is countered, many alternatives remain to explore.

The third factor relates to the tradeoffs inherent in both generation and detection. Watermarking techniques robust against evasion typically impose costs on either generation quality or computational efficiency, creating incentives to use less robust but more efficient approaches. Detection techniques that minimize false positives typically sacrifice sensitivity, missing some AI-generated content. These inherent tradeoffs mean that perfect detection without costs or limitations remains unattainable, leaving space for evasion.

The fourth factor involves the fundamental indistinguishability between high-quality AI-generated content and authentic human-created content. As generation systems improve, the statistical properties of synthetic content converge with those of authentic content. At the limit, if AI systems could perfectly mimic the full distribution of human-created content, detection based on content properties alone would become theoretically impossible. While this perfect mimicry has not been achieved and may face fundamental obstacles, the trend toward higher generation quality makes detection progressively more challenging.

The fifth factor concerns the active nature of adversaries who continuously adapt and innovate specifically to evade detection. Unlike passive challenges where a solution once found remains effective, adversarial challenges involve intelligent opponents who study defenses and design attacks to exploit weaknesses. This active adaptation ensures that static defenses eventually fail, requiring continuous innovation to maintain effectiveness.

Embracing Continuous Adaptation

Given that perfect permanent solutions to watermarking and detection challenges remain elusive, the practical approach must embrace continuous adaptation rather than seeking to achieve final victory. This adaptation mindset shapes strategy, resource allocation, system design, and expectations in ways that differ fundamentally from approaches assuming stable solutions are achievable.

Embracing continuous adaptation means treating watermarking and detection as ongoing processes requiring sustained investment rather than one-time projects that produce lasting solutions. Organizations and institutions relying on detection capabilities must plan for continuous monitoring, research, and system updates rather than assuming that initial deployment solves the problem permanently. Budget and staffing decisions should reflect the ongoing nature of the challenge rather than treating detection as solved once initial systems are in place.

System design should anticipate the need for evolution, building in flexibility and modularity that enable updates to detection techniques without requiring complete system redesign. Rather than hardcoding specific watermarking or detection approaches, systems should employ architectures that allow techniques to be updated as adversarial innovation proceeds. The ability to rapidly deploy new detection approaches when existing ones are evaded proves essential for maintaining effectiveness in adversarial environments.

Measurement and monitoring systems should track not just current detection performance but also trends indicating emerging evasion techniques or degrading effectiveness. Early warning systems that detect when evasion attempts increase or when detection reliability decreases enable proactive responses before problems become severe. Continuous collection of ground truth data about actual AI generation and human creation provides the information necessary for ongoing refinement of detection approaches.

Collaboration between researchers, platform operators, policymakers, and other stakeholders becomes essential when facing challenges requiring continuous adaptation. No single organization can sustain the level of innovation necessary to keep pace with adversarial developments. Sharing information about new evasion techniques, coordinating research on countermeasures, and collectively advancing the state of detection capability prove more effective than isolated efforts. However, this collaboration must be balanced against the risk that information sharing could also benefit adversaries seeking to evade detection.

The mindset of continuous adaptation also requires honesty about limitations and uncertainty. Rather than claiming that detection systems provide perfect reliability, communication should acknowledge their limitations, describe conditions under which they work well or poorly, and explain that effectiveness may degrade over time as evasion techniques evolve. This honesty builds more appropriate trust than overpromising capabilities that cannot be sustained.

Implications for Different Stakeholders

The recognition that watermarking and detection exist in perpetual adversarial co-evolution carries different implications for various stakeholders involved with AI-generated content, each needing to adapt their expectations and approaches to this reality.

For platform operators seeking to identify and manage AI-generated content, the continuous adaptation reality means that content moderation systems must evolve constantly rather than relying on static detection capabilities. Investment in detection technology must be sustained over time rather than treated as one-time expense. Policies about AI-generated content must acknowledge that detection is imperfect and may degrade, building in appropriate flexibility and human judgment rather than relying purely on automated systems.

For AI developers creating generation systems, the adversarial dynamics create ongoing tension between improving generation quality and cooperating with detection efforts. Responsible developers must navigate the reality that making content generation more difficult to detect could enable harmful uses, while implementing robust watermarking might disadvantage legitimate users of their systems. Finding appropriate balances requires engaging with broader ecosystem of stakeholders rather than optimizing purely for generation capability.

For policymakers considering regulations around AI-generated content and its identification, the absence of perfect detection solutions means that policies cannot rely on assumed perfect distinguishability of synthetic and authentic content. Regulations must account for false positives where authentic content is incorrectly flagged as AI-generated, false negatives where AI content evades detection, and evolving detection reliability over time. Prescriptive mandates for specific detection approaches risk becoming obsolete as adversarial innovation proceeds, suggesting more flexible regulatory frameworks that adapt to technological evolution.

For researchers working on either generation or detection, the adversarial dynamics create both challenges and opportunities. The continuous need for innovation ensures ongoing research relevance and funding, but also creates pressure to advance capabilities with full recognition that any advance will eventually be countered. Ethical considerations about responsible disclosure and dual-use implications of research become particularly salient in adversarial contexts where innovations in either direction can be weaponized.

For users of AI systems and consumers of content, the imperfect and evolving state of detection means that strong certainty about content provenance often remains unattainable. Building appropriate media literacy that accounts for this uncertainty, developing healthy skepticism about content origins, and understanding the limitations of automated verification systems all become important capabilities for navigating environments where synthetic and authentic content intermingle.

The Long View on Adversarial Innovation

Taking a long view on the adversarial innovation dynamic between AI generation and watermarking detection reveals patterns and principles that can guide more effective engagement with these challenges. While the specific technical details of generation and detection approaches will continue evolving in ways we cannot predict, the broader dynamics follow patterns that prove more stable and predictable.

The history of adversarial innovation in other domains provides instructive parallels. In cybersecurity, the dynamic between attackers and defenders has persisted for decades with neither side achieving permanent dominance. Successful security strategies embrace continuous adaptation, defense in depth with multiple layers of protection, and acceptance that some attacks will succeed requiring response and recovery capabilities. Similar patterns appear in fraud detection, spam filtering, and other domains involving adversarial dynamics.

These parallels suggest that effective long-term approaches to AI content detection will similarly require multiple complementary strategies rather than relying on any single technique, continuous monitoring and adaptation as evasion techniques evolve, layered defenses where evasion of one technique does not compromise all detection capability, and acceptance that perfect detection remains unattainable requiring graceful handling of both false positives and false negatives.

The co-evolutionary dynamic also creates opportunities alongside challenges. The continuous need for innovation drives research that advances understanding of both generation and detection, producing insights applicable beyond the specific adversarial context. The pressure to make watermarking robust and difficult to evade drives innovation in signal processing and cryptography. The challenge of detecting subtle differences between synthetic and authentic content advances understanding of statistical patterns in different types of media.

Preparing for an Uncertain Future

The symbiotic co-evolution of generative AI systems and watermarking detection technologies defines a future characterized by uncertainty about specific technical capabilities at any given moment but predictable patterns in the broader dynamics. Preparing for this future requires embracing the reality of continuous change, building systems and institutions that can adapt rather than assuming stability, and maintaining appropriate humility about the limitations of both generation and detection capabilities.

The practical wisdom for stakeholders across the ecosystem involves treating both generation capability and detection reliability as continuously evolving rather than fixed, investing in monitoring and adaptation rather than assuming that current solutions will remain adequate, building flexibility into systems, policies, and strategies to accommodate evolution, and maintaining realistic expectations about what detection can and cannot achieve.

The relationship between AI generation and watermarking detection exemplifies broader patterns in technology where adversarial dynamics drive perpetual innovation. Understanding these patterns, accepting the absence of perfect permanent solutions, and embracing approaches based on continuous adaptation positions individuals, organizations, and society to navigate more effectively the complex challenges arising as synthetic and authentic content become increasingly difficult to distinguish. The future of watermarking is indeed tied to the future of AI itself, with the two co-evolving in an ongoing dance of innovation and counter-innovation that will continue reshaping both capabilities in ways that remain difficult to predict but follow understandable broader patterns.

Conclusion:

I believe that AI watermarking, despite its challenges and limitations, holds immense and unavoidable potential for building trust and transparency. We cannot put the generative AI “genie” back in the bottle. The technology is here, and its capabilities will only grow. Therefore, our only path forward is to build a robust ecosystem of tools for safety, accountability, and trust. Watermarking is the most promising technical foundation for that ecosystem. By enabling the identification of AI-generated content, it can combat misinformation, protect intellectual property, and promote the ethical use of AI. For me, the most exciting aspect is how it empowers individuals to make informed decisions about the content they interact with. While this is true, significant challenges remain. This is why ongoing research, open collaboration, and difficult ethical conversations are so crucial.