A new digital watermarking technique for MIT CSAIL seeks to prevent unauthorized image edits by malicious AI…

  • Itty53@kbin.social
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    1 year ago

    Really wanna see how it handles the standard Photoshop touch ups. It’s not like the news media has never altered photos to solicit a skewed perception.

    • CaptainPatent@kbin.social
      link
      fedilink
      arrow-up
      0
      ·
      1 year ago

      The big problem is AI will (eventually) “see” things as a human does so even in the case that these MIT researchers are able to insert nearly invisible artifacts that fool AI into thinking the edges are different than they actually are, a sufficiently large training set will allow the AI to see that the color borders are more important than artifact borders…

      Which will allow AI to bypass this type of watermarking.

      • Itty53@kbin.social
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        1 year ago

        I really hate the label AI. They’re data models, not intelligence - artificial or otherwise. It’s PAI. Pseudo Artificial Intelligence, which we’ve had since the 80s.

        The thing is that these data models are, in the end, fed to algorithms to provide output. That being the case it’s a mathematical certainty that it can be reversed and thus, shown to be from such an algorithm. Watermark or not, if an algorithm makes a result, then you can deduce the algorithm from a given set of it’s results.

        It wouldn’t be able to meaningfully distinguish 4’33" from silence though. Nor could it determine a flat white image wasn’t made by an algorithm.

        I think what we’re really demonstrating in all this is just exactly how algorithmically human beings think already. Something psychology has been talking about for a longer time still.

        • BraveSirZaphod@kbin.social
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          1 year ago

          It wouldn’t be able to meaningfully distinguish 4’33" from silence though.

          Nor could a human though, no? There’s obviously a lot of metadata about 4’33" that makes it what it is - namely that it is a published work that is performed - but an actual recording of it is silence, so I’m not really sure what this apparent limitation that you’re talking about really is.

          Edit: and an AI could observe and analyze that metadata just as much as a human could, provided it has access to it.