A judge in Washington state has blocked video evidence that’s been “AI-enhanced” from being submitted in a triple murder trial. And that’s a good thing, given the fact that too many people seem to think applying an AI filter can give them access to secret visual data.

  • Bobby Turkalino@lemmy.yachts
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    109
    ·
    10 个月前

    Everyone uses the word “hallucinate” when describing visual AI because it’s normie-friendly and cool sounding, but the results are a product of math. Very complex math, yes, but computers aren’t taking drugs and randomly pooping out images because computers can’t do anything truly random.

    You know what else uses math? Basically every image modification algorithm, including resizing. I wonder how this judge would feel about viewing a 720p video on a 4k courtroom TV because “hallucination” takes place in that case too.

    • Downcount@lemmy.world
      link
      fedilink
      English
      arrow-up
      77
      arrow-down
      2
      ·
      10 个月前

      There is a huge difference between interpolating pixels and inserting whole objects into pictures.

      • Bobby Turkalino@lemmy.yachts
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        42
        ·
        10 个月前

        Both insert pixels that didn’t exist before, so where do we draw the line of how much of that is acceptable?

        • Downcount@lemmy.world
          link
          fedilink
          English
          arrow-up
          56
          arrow-down
          3
          ·
          edit-2
          10 个月前

          Look it this way: If you have an unreadable licence plate because of low resolution, interpolating won’t make it readable (as long as we didn’t switch to a CSI universe). An AI, on the other hand, could just “invent” (I know, I know, normy speak in your eyes) a readable one.

          You will draw yourself the line when you get your first ticket for speeding, when it wasn’t your car.

          • Natanael@slrpnk.net
            link
            fedilink
            English
            arrow-up
            8
            ·
            edit-2
            10 个月前

            License plates is an interesting case because with a known set of visual symbols (known fonts used by approved plate issuers) you can often accurately deblur even very very blurry text (but not with AI algorithms, but rather by modeling the blur of the cameras and the unique blur gradients this results in for each letter). It does require a certain minimum pixel resolution of the letters to guarantee unambiguity though.

          • Bobby Turkalino@lemmy.yachts
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            20
            ·
            10 个月前

            Interesting example, because tickets issued by automated cameras aren’t enforced in most places in the US. You can safely ignore those tickets and the police won’t do anything about it because they know how faulty these systems are and most of the cameras are owned by private companies anyway.

            “Readable” is a subjective matter of interpretation, so again, I’m confused on how exactly you’re distinguishing good & pure fictional pixels from bad & evil fictional pixels

            • Downcount@lemmy.world
              link
              fedilink
              English
              arrow-up
              22
              arrow-down
              1
              ·
              10 个月前

              Being tickets enforced or not doesn’t change my argumentation nor invalidates it.

              You are acting stubborn and childish. Everything there was to say has been said. If you still think you are right, do it, as you are not able or willing to understand. Let me be clear: I think you are trolling and I’m not in any mood to participate in this anymore.

              • Bobby Turkalino@lemmy.yachts
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                21
                ·
                10 个月前

                Sorry, it’s just that I work in a field where making distinctions is based on math and/or logic, while you’re making a distinction between AI- and non-AI-based image interpolation based on opinion and subjective observation

                • pm_me_ur_thoughts@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  11
                  ·
                  10 个月前

                  Okay, I’m not disagreeing with you about the fact that its all math.

                  However, interpolation or pixels is simple math. AI generated is complex math and is only as good as its training data.

                  The licence example is a good one. In interpolation, it’ll just find some average, midpoint, etc and fill the pixel. In AI gen, if the training set had your number plate 999 times in a set of 1000, it will generate your numberplate no matter whose plate you input. to use it as evidence would need it to be far more deterministic than the probabilistic nature of AI gen content.

            • abhibeckert@lemmy.world
              link
              fedilink
              English
              arrow-up
              8
              ·
              edit-2
              10 个月前

              You can safely ignore those tickets and the police won’t do anything

              Wait what? No.

              It’s entirely possible if you ignore the ticket, a human might review it and find there’s insufficient evidence. But if, for example, you ran a red light and they have a photo that shows your number plate and your face… then you don’t want to ignore that ticket. And they generally take multiple photos, so even if the one you received on the ticket doesn’t identify you, that doesn’t mean you’re safe.

              When automated infringement systems were brand new the cameras were low quality / poorly installed / didn’t gather evidence necessary to win a court challenge… getting tickets overturned was so easy they didn’t even bother taking it to court. But it’s not that easy now, they have picked up their game and are continuing to improve the technology.

              Also - if you claim someone else was driving your car, and then they prove in court that you were driving… congratulations, your slap on the wrist fine is now a much more serious matter.

        • Blackmist
          link
          fedilink
          English
          arrow-up
          18
          ·
          edit-2
          10 个月前

          I mean we “invent” pixels anyway for pretty much all digital photography based on Bayer filters.

          But the answer is linear interpolation. That’s where we draw the line. We have to be able to point to a line of code and say where the data came from, rather than a giant blob of image data that could contain anything.

        • Catoblepas@lemmy.blahaj.zone
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          10 个月前

          What’s your bank account information? I’m either going to add or subtract a lot of money from it. Both alter your account balance so you should be fine with either right?

    • Catoblepas@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      38
      arrow-down
      1
      ·
      10 个月前

      Has this argument ever worked on anyone who has ever touched a digital camera? “Resizing video is just like running it through AI to invent details that didn’t exist in the original image”?

      “It uses math” isn’t the complaint and I’m pretty sure you know that.

    • abhibeckert@lemmy.world
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      1
      ·
      edit-2
      10 个月前

      computers aren’t taking drugs and randomly pooping out images

      Sure, no drugs involved, but they are running a statistically proven random number generator and using that (along with non-random data) to generate the image.

      The result is this - ask for the same image, get two different images — similar, but clearly not the same person - sisters or cousins perhaps… but nowhere near usable as evidence in court:

      • Gabu@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        18
        ·
        10 个月前

        Tell me you don’t know shit about AI without telling me you don’t know shit. You can easily reproduce the exact same image by defining the starting seed and constraining the network to a specific sequence of operations.

        • Natanael@slrpnk.net
          link
          fedilink
          English
          arrow-up
          9
          ·
          10 个月前

          But if you don’t do that then the ML engine doesn’t have the introspective capability to realize it failed to recreate an image

          • Gabu@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            13
            ·
            10 个月前

            And if you take your eyes off of their sockets you can no longer see. That’s a meaningless statement.

            • blind3rdeye@lemm.ee
              link
              fedilink
              English
              arrow-up
              3
              ·
              10 个月前

              The point is that the AI ‘enhanced’ photos have nice clear details that are randomly produced, and thus should not be relied on. Are you suggesting that we can work around that problem by choosing a random seed manually? Do you think that solves the problem?

    • Malfeasant@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      arrow-down
      2
      ·
      10 个月前

      computers can’t do anything truly random.

      Technically incorrect - computers can be supplied with sources of entropy, so while it’s true that they will produce the same output given identical inputs, it is in practice quite possible to ensure that they do not receive identical inputs if you don’t want them to.

      • Hackerman_uwu@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        10 个月前

        IIRC there was a random number generator website where the machine was hookup up to a potato or some shit.

    • Kedly@lemm.ee
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      2
      ·
      10 个月前

      Bud, hallucinate is a perfect term for the shit AI creates because it doesnt understand reality, regardless if math is creating that hallucination or not