(despite david simon being a shitlib)

  • Llituro [he/him, they/them]@hexbear.net
    link
    fedilink
    English
    arrow-up
    27
    ·
    5 months ago

    the only use of ai i think is probably remotely useful is programmers using it to help write new code. not people who aren’t experienced at software development mind you, they don’t get too much of chatgpt, but someone that knows what they’re doing with copilot to copy-paste someone’s completely correct implementation, that seems useful. at least to people i’ve talked to.

    • SoyViking [he/him]@hexbear.net
      link
      fedilink
      English
      arrow-up
      31
      ·
      5 months ago

      It is very useful for coding because that is one of the few places where unoriginal repetitive solutions are often desirable. But even with coding you have to know what to tell the LLM to do and you have to be able to read and understand the output to make sure it works as intended.

      LLM’s are a useful too for programmers to automate repetitive tasks but it is nowhere near bearing able to produce usable applications by itself. I am not worried that I’ll be replaced by a robot anytime soon.

      Those who should be worried about their jobs are people in places like customer support or government services directed at people who doesn’t matter to the ruling class. In these cases the powers that be have little holding them from replacing human interactions with significantly worse interactions with a LLM. Nobody important gives a shit if some schmuck can’t cancel their cable subscription or gets their employment benefits cut because the computer had a hiccup.

    • macerated_baby_presidents [he/him]@hexbear.net
      link
      fedilink
      English
      arrow-up
      22
      ·
      edit-2
      5 months ago

      IMO no, for two reasons:

      • reading code is harder than writing it. If the AI writes you a standard implementation, you still have to read it to make sure it’s correct. So that’s more work than just doing it yourself
      • AI will produce code that looks right. Since it can’t understand anything that’s all it does, next most likely token == most correct-looking solution. But when the obvious solution is not the right one, you now have deceptively incorrect code, specifically and solely designed to look correct.

      I’ve never used Copilot myself but pair programmed with someone who used it, and it seemed like he spent more time messing with the output than it would have taken to write it himself.

    • Cysioland@lemmygrad.ml
      link
      fedilink
      English
      arrow-up
      7
      ·
      5 months ago

      I use JetBrains “local LLM” thingy and it’s good at suggesting the very obvious, trivial code that I would write anyway, so it just saves me keystrokes

    • gaycomputeruser [she/her]@hexbear.net
      link
      fedilink
      English
      arrow-up
      5
      ·
      5 months ago

      It’s clearly become a crutch for some programmers. I remember talking to someone who does ai research and openly admitted that most of the people in their lab couldn’t code and that the outputs from chatgpt where sufficient to do their work.