• linearchaos@lemmy.world
    link
    fedilink
    English
    arrow-up
    11
    arrow-down
    4
    ·
    5 hours ago

    We need to stop viewing it as artificial intelligence. The parts that are worth money are just more advanced versions of machine learning.

    Being able to assimilate a few dozen textbooks and pass a bar exam is a neat parlor trick, but it is still just a parlor trick.

    Unfortunately probably the biggest thing to come out of it will be the marketing aspect. If they spend enough money to train small models on our wants and likes it will give them tremendous amounts of return.

    The key to using it in a financially successful manner is finding problems that fit the bill. Training costs are fairly high, quality content generation is also rather expensive. There are sticky problems around training it from non-free data. Whatever you’re going to use it for either needs to have a significant enough advantage to make the cost of training /data worth it.

    I still think we’re eventually going to see education rise. The existing tools for small content generation adobe’s use of it to fill in small areas is leaps and bounds better than the old content aware patches. We’ve been using it for ages for speech recognition and speech generation. From there it’s relatively good at helper roles. Minor application development, copy editing, maybe some VFX generation eventually. Things where you still need a talented individual to oversee it but it can help lessen the workload.

    There are lots of places where it’s being used where I think it’s a particularly poor fit. AI help desk chatbots, IVR scenarios, It says brain dead as the original phone trees and flow charts that we’ve been following for decades.

    • Eheran@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      6
      ·
      2 hours ago

      If GPT4o is still not what you would call AI, then what is? You can have conversations with it, the Turing test is completely irrelevant all of the sudden.

      • zarkanian@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        3
        ·
        24 minutes ago

        Hasn’t the Turing Test been irrelevant for a while now? Even before the new AI boom?

        Artificial intelligence is a moving target. Every time a goal gets reached, they just move the goalposts, because “well, obviously this isn’t real intelligence”.

      • Buddahriffic@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        arrow-down
        2
        ·
        2 hours ago

        It’s a massive text predictor. It doesn’t solve problems, it applies patterns based on correlations it picked up during training. If someone talked about your topic online, it has been trained on those conversations. If a topic has two sides that don’t agree, chat gpt might respond in a way that is biased towards one side or the other and you can easily get it to “switch” to the other side with follow up prompts.

        For what would be considered AI, think of the star trek computer or Data. The Star Trek computer could create simulations of warp core behaviour to push frontiers of knowledge or characters smart enough to defeat its own safeties (frankly, the computer was such a deus ex machina kinda thing that it was hard to suspend disbelief at times, like why did they even have humans doing the problem solving with computers that capable?). Data wouldn’t get confused about whether any counties in Africa start with K.

        I don’t think the Turing test is an effective means of determining intelligence anyways. It came from a time when a conversational computer was barely thinkable. But I wouldn’t even say chat gpt is there yet, since you can tell if you ask it the right things. It is very useful, don’t get me wrong, like a very powerful search engine. But it’s not intelligent.

        • Eheran@lemmy.world
          link
          fedilink
          English
          arrow-up
          7
          arrow-down
          3
          ·
          2 hours ago

          What of what you say does not apply to humans? They apply patterns of behavior in response to some input. Picked up by learning them. Including people talking online. They are always biased on some way. Some will acknowledge their bias and change it if you give them context.

          GPT can literally create simulations. I have used it to do exactly that, specifically for 2D heat conducting with coupled mass transport and reaction kinetics.

          • Buddahriffic@lemmy.world
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            60 minutes ago

            Yeah, it does do some very human-like things, but it’s still missing some important parts.

            It’s kinda like using a textbook for problem solving. It’s great at helping you solve instances of problems that have already been solved, but you won’t likely find the next big advancement in that field in a textbook.

            Newton realized masses attracted each other, and through experimentation, came up with his laws of classical physics.

            Einstein took the idea that the speed of light always seems to be the same despite relative motion to come up with special relativity, then realized that space-time itself was a physical thing that could be interacted with rather than just a medium, plus came up with field equations that were used to predict things like black holes before anyone had any kind of notion that they were real things.

            Chat gpt is incapable of things like that. And sure, many humans never do anything like that, some might not even be capable even if they were motivated and had the right supports to try. But many humans do solve problems that they’ve never seen before. There’s big names in academia but so many more that don’t get famous but still push the boundaries of human knowledge, creatively solving problems and answering questions every day.

            I wouldn’t be surprised if an LLM is a piece of general AI if or when it comes, but there will be other parts that are currently missing. We don’t even know what consciousness is, let alone if any of our hardware is capable of creating/hosting one.

            • EatATaco@lemm.ee
              link
              fedilink
              English
              arrow-up
              1
              ·
              13 minutes ago

              I listened to a podcast (This American Life, IIRC), where some researchers were talking about their efforts to determine whether or not AI could reason. One test they did was asking it to stack a random set of items (one it wouldn’t have come across in any data set, plank of wood, 12 eggs, a book, a bottle, and a nail. . .probably some other things too) in a stable way. With chat gpt 3, it basically just (as you would expect from a pure text predictor) said to put one object on top of another, no way would it be stable.

              However, with gpt 4, it basically said to put the wood down, and place the eggs in a 3 x 4 grid with the book on top (to stop them from rolling away), and then with the bottle on top of that, with the nail (even noting you have to put the head side down because you couldn’t make it stable with the point down). It was certainly something that could work, and it was a novel solution.

              Now I’m not saying this proves it can think, but I think this “well it’s just a text predictor” kind of hand-waves away the question. It also begs the question, and based on how often I hear people parroting the same exact arguments against AI thinking, I wonder how much we are simply just “text predictors.”

      • Cryophilia@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        6
        ·
        2 hours ago

        I can write a program that just replies “yes” to everything you say and you can have a conversation with that. Is that program AI?

        “AI isn’t really AI and no one ever thought that AI was actually AI so it doesn’t matter if we call it AI” is the funniest level of tech bro cope these days.