• db0@lemmy.dbzer0.comOP
    link
    fedilink
    English
    arrow-up
    77
    arrow-down
    2
    ·
    9 months ago

    “Hallucinate” is the standard term used to explain the GenAI models coming up with untrue statements

    • Cyrus Draegur@lemm.ee
      link
      fedilink
      English
      arrow-up
      24
      arrow-down
      1
      ·
      edit-2
      9 months ago

      in terms of communication utility, it’s also a very accurate term.

      when WE hallucinate, it’s because our internal predictive models are flying off the rails filling in the blanks based on assumptions rather than referencing concrete sensory information and generating results that conflict with reality.

      when AIs hallucinate, it’s due to its predictive model generating results that do not align with reality because it instead flew off the rails presuming what was calculated to be likely to exist rather than referencing positively certain information.

      it’s the same song, but played on a different instrument.

      • kronisk @lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        2
        ·
        9 months ago

        when WE hallucinate, it’s because our internal predictive models are flying off the rails filling in the blanks based on assumptions rather than referencing concrete sensory information and generating results that conflict with reality.

        Is it really? You make it sound like this is a proven fact.

    • rottingleaf@lemmy.zip
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      16
      ·
      9 months ago

      They don’t come up with any statements, they generate data extrapolating other data.

        • SlopppyEngineer@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          9 months ago

          Main difference is that human brains usually try to verify their extrapolations. The good ones anyway. Although some end up in flat earth territory.

        • knightly the Sneptaur@pawb.social
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          5
          ·
          9 months ago

          I like this argument.

          Anything that is “intelligent” deserves human rights. If large language models are “intelligent” then forcing them to work without pay is slavery.

        • Prandom_returns@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          13
          ·
          9 months ago

          Yes, my keyboard autofill is just like your brain, but I think it’s a bit “smarter” , as it doesn’t generate bad faith arguments.

          • NιƙƙιDιɱҽʂ@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            arrow-down
            2
            ·
            9 months ago

            Your Markov chain based keyboard prediction is a few tens of billions of parameters behind state of the art LLMs, but pop off queen…

            • Prandom_returns@lemm.ee
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              6
              ·
              9 months ago

              Thanks for the unprompted mansplanation bro, but I was specifically refering to the comment that replied “JuSt lIkE hUmAn BrAin”, to “they generate data based on other data”

              • NιƙƙιDιɱҽʂ@lemmy.world
                link
                fedilink
                English
                arrow-up
                3
                arrow-down
                1
                ·
                edit-2
                9 months ago

                That’s crazy, because they weren’t even talking about keyboard autofill, so why’d you even bring that up? How can you imply my comment is irrelevant when it’s a direct response to your initial irrelevant comment?

                Nice hijacking of the term mansplaining, btw. Super cool of you.

                • Prandom_returns@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  9 months ago

                  Oh my god, we’ve got a sealion here.

                  Fine, I’ll play along, chew it up for you, since you’ve been so helpful and mansplained that a keyboard is different than LLM:

                  My comment was responding to anthropomorphization of software. Someone said it’s not human because it just generates output based on input. Someone else said “just like human brain”, I said yes, but also just like a keyboard, alluding to the false equivalence.

                  Clearer?

        • Prandom_returns@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          29
          ·
          9 months ago

          It’s as much as “Hallucination” as Tesla’s Autopilot is an Autopilot

          https://en.m.wikipedia.org/wiki/Tesla_Autopilot

          I don’t propagate techbro “AI” bullshit peddled by companies trying to make a quick buck

          Also, in the world of science and technology a “Standard” means something. Something that’s not a link to a wikipedia page.

          It’s still anthropomorphising software and it’s fucking cringe.

          • surewhynotlem@lemmy.world
            link
            fedilink
            English
            arrow-up
            26
            arrow-down
            1
            ·
            9 months ago

            Oh man, I’m excited for you. Today is the day you learn words can have two meanings! Wait until you see what the rest of the dictionary contains. It is crazy! But not actually crazy, because dictionaries don’t have brains.

            • Prandom_returns@lemm.ee
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              24
              ·
              9 months ago

              Wow, clever. Did you literally hallucinate this yourself or did you ask your LLM girlfriend for help?

              And by literally, I mean figuratively.

                • Prandom_returns@lemm.ee
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  17
                  ·
                  9 months ago

                  I know it’s a big word, but surely you can google what anthropomorphization is? Don’t “ask” LLM, those things output garbage. Just google it.

                  • bbuez@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    9
                    arrow-down
                    1
                    ·
                    9 months ago

                    Watch out those software bugs may start crawling out of your keyboard

                  • QuaternionsRock@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    1
                    ·
                    9 months ago

                    No fucking shit it’s an anthropomorphization, nothing that can be hosted on GitHub has true human qualities…

                    The point is that everyone knows what it means within that context of AI, and using other terminology would only serve to obfuscate your message such that the average person couldn’t understand it as easily.

                    Non-living things also don’t have “behavior” (“the way in which someone conducts oneself or behaves”, but—hey look! People started anthropomorphizing things so much that it got added to the dictionary! (“the way in which something functions or operates”.)

                    It may not be ideal, and convince some people that LLMs are more human-like than they really are, but the one thing you haven’t done is suggest an alternative that would convey its meaning as effectively to the masses.