Reverse engineering Gemini's SynthID detection

(github.com)

87 points | by _tk_ 3 hours ago

11 comments

  • coppsilgold 1 hour ago
    Inserting an undetectable 1-bit watermark into a multi megapixel image is not particularly difficult.

    If you assume competence from Google, they probably have two different watermarks. A sloppy one they offer an online oracle for and one they keep in reserve for themselves (and law enforcement requests).

    Also given that it's Google we are dealing with here, they probably save every single image generated (or at least its neural hash) and tie it to your account in their database.

  • Tiberium 1 hour ago
    Seems like a very low-quality AI-assisted research repo, and it doesn't even properly test against Google's own SynthID detector. It's not hard at all (with some LLM assistance, for example) to reverse-engineer network requests to be able to do SynthID detection without a browser instance or Gemini access, and then you'd have a ground truth.
    • ddtaylor 57 minutes ago
      I read a lot of comments on HN that say something is not hard, yet don't provide a POC of their own or link to research they have knowledge of.

      I also read a lot of comments on HN that start by attacking the source of the information, such as saying it was AI assisted, instead of the actual merits of the work.

      The HN community is becoming curmudgeonly and using AI tooling as the justification.

      • love2read 1 minute ago
        becoming? under most posts that even in passing mention using AI tools there are multiple people raising their noses talking about how much they hate AI use
  • armanj 2 hours ago
    kinda ironic you can clearly see signs of Claude, as it shows misaligning table walls in the readme doc
    • rafram 2 hours ago
      Parenthesized, comma-separated lists with no “and” is an even stronger tell. Claude loves those.
    • TacticalCoder 1 hour ago
      > kinda ironic you can clearly see signs of Claude, as it shows misaligning table walls in the readme doc

      This one is such a gigantic clusterfuck... They're mimicking ASCII tables using Unicode chars of varying length and, at times, there's also an off-by-one error. But the model (not Claude, but the model underneath it) is capable of generating ASCII tables.

      P.S: I saw the future... The year is 2037 and we've got Unicode tables still not properly aligned.

    • dgellow 1 hour ago
      I mean, just reading the readme content it is pretty obvious it is Claude
    • huflungdung 1 hour ago
      [dead]
  • khernandezrt 2 hours ago
    Ok i get that eventually someone was gonna do this but why would we want to purposely remove one of the only ways of detecting if an image is ai generated or not...?
    • akersten 35 minutes ago
      Fundamentally it's a fuzzy signal and people shouldn't rely on it. The general public does not understand Boolean logic (oh, so the SynthID is not there, therefore this image is real). The sooner AI watermarking faces its deserved farcical demise the better.

      Also something about how AI is not special and we haven't added or needed invisible watermarks for other ways media can be manipulated deceptively since time immemorial, but that's less of a practical argument and more of a philosophical one.

      • StarlaAtNight 9 minutes ago
        I’m not very well read on the topic and you seen to take a strong “con” stance. Curious to hear why you think it deserves such a demise
    • lokar 2 hours ago
      It was always going to be available to some people, but not everyone would know or believe that. Now they will.
      • subscribed 1 hour ago
        More likely than not it would be used to deanonymise the author.

        So it's a "no" by default.

    • raincole 2 hours ago
      Uh... you can do this pretty easily since day 1. Just use Stable Diffusion with a low denoising strength. This repo presents an even less destructive way[0], but it has always been very easy to hide that an image is generated by Nano Banana.

      [0]: if it does what it claims to do. I didn't verify. Given how much AI writing in the README my hunch is that this doesn't work better than simple denoising.

  • M4v3R 2 hours ago
    SynthID is visible in some generations (areas with a lot of edges, or text), I wonder if this would make them look better.
  • doctorpangloss 1 hour ago
    Okay... this tests its own ability to remove the watermark against its own detector. It doesn't test against Gemini's SynthID app. So it does nothing...
  • kelsey98765431 2 hours ago
    if you downscale then upscale it removes the watermark
  • sodacanner 1 hour ago
    I don't understand all the handwringing. If it's this easy to remove SynthID from an AI-generated image then it wasn't a good solution in the first place.
    • raincole 1 hour ago
      There is no solution. I don't know why people discuss this subject as if there is a technical solution. As if there are fairies or souls hidden in the pixels that help us tell what is AI generated and what is not.
      • DonsDiscountGas 1 hour ago
        If you want to make an AI generated image but don't want other people to know that it's AI, the most obvious solution is to not use Gemini. Synth ID is watermarking. It's only ever going to be useful to good actors, who want an AI generated image and aren't trying to hide the fact that it's AI generated.
        • dummydummy1234 53 minutes ago
          Never underestimate that people are lazy.
      • levocardia 1 hour ago
        Sure there is a solution, you are just looking at it the wrong way. Make non-AI images provably unaltered with signed keys from the device (e.g. the camera) that took it.
        • jfim 39 minutes ago
          That's pretty much impossible though.

          One workflow that some artists use is that they draw with ink on paper, scan, and then digitally color. Nothing prevents someone from generating line art using generative AI, printing it, scanning it, and coloring it.

          And what if someone just copy pastes something into Photoshop or imports layers? That's what you'd do for composites that mix multiple images together. Can one copy paste screenshots into a multi layer composition or is that verboten and taints the final image?

          And what about multi program workflows? Let's say I import a photo, denoise it in DxO, retouch in affinity photo, resize programmatically using image magick, and use pngcrush to optimize it, what metadata is left at the end?

        • raincole 42 minutes ago
          If the premise is that everyone would just agree on the same protocol, I have an even more unbreakable solution: every image has to be upload to a blockchain the moment it is (claimed to be) created. Otherwise it's AI.

          If only everyone just agrees with me.

        • Diggsey 1 hour ago
          Which works for about 5 minutes until someone leaks a manufacturer's private key or extracts it from a device...
        • IncreasePosts 1 hour ago
          How many minutes do you think it would take before someone figured out how to crack that?
          • subscribed 59 minutes ago
            On Pixels and iPhones it would be impossible since they have actually secure hardware that could both hold the keys and sign/verify the image.
            • IncreasePosts 49 minutes ago
              The camera module sits outside the secure area, meaning it would need to send data in to be signed. How does the phone know that it's getting legitimate data from the camera module, or data someone else is just piping in? Also, you could probably get a fairly high quality image by just taking a photo of something AI generated in the right lighting conditions.
      • sodacanner 1 hour ago
        Sure, and things like this help drive home that SynthID wasn't a solution at all.
    • rustyhancock 1 hour ago
      Yes. This kind of project needs aggressive red teaming, it leads to better products and we need excellent products in this space.

      This project proves what red teaming was in place wasn't good enough.

  • andrewmcwatters 2 hours ago
    > We're actively collecting pure black and pure white images generated by Nano Banana Pro to improve multi-resolution watermark extraction.

    Oh hey, neat. I mentioned this specific method of extracting SynthID a while back.[1]

    Glad to see someone take it up.

    [1]: https://news.ycombinator.com/item?id=47169146#47169767

    • raphman 2 hours ago
      FWIW, I had Nano Banana create pure white/black images in February, and there was no recognizable watermark in them (all pixels really were #ffffff / #000000 IIRC).

      Meta: your comment was marked [dead], like a few other constructive comments I saw in recent days. Not sure why.

      • andrewmcwatters 1 hour ago
        I suspect they strip the SynthID for these specific cases to prevent exfiltration of the steganography.

        I appreciate you pointing it out, but this account is banned. Thank you for vouching though!

  • refulgentis 2 hours ago
    [flagged]
    • jonshariat 2 hours ago
      Agreed. This isn't punk this just helps the bad guys. Society needs to know what content is AI generated and what is not.
      • recursive 1 hour ago
        This was never going to be a reliable way to do it. It's basically the evil bit . It only works for as long as everyone is making a good-faith effort to follow the convention. But the bad guys do not do that.
      • SR2Z 2 hours ago
        If that's the case, society will inevitably be disappointed.

        There are already ten million AI image generators, the overwhelming majority of which do not watermark their outputs. Google auto-inserting them is nice, but ultimately this kind of tool to remove them will inevitably be widespread.

      • charcircuit 1 hour ago
        It really doesn't need such capability. Nor does it need the capability to know what human generated it either.