The New York Times blocks OpenAI’s web crawler::The New York Times has officially blocked GPTBot, OpenAI’s web crawler. The outlet’s robot.txt page specifically disallows GPTBot, preventing OpenAI from scraping content from its website to train AI models.

  • porkins@lemmy.basedcount.com
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    46
    ·
    edit-2
    1 year ago

    This goes against everything that the NYT preaches in terms of saying that the press is under attack and needs to be protected. AI consumption of news content makes the news more accessible. Their paid articles don’t overlap with what ChatGPT is doing. This is really a bunch of old people getting butt hurt about tech they don’t fully understand.

    • Kachilde@lemmy.world
      link
      fedilink
      English
      arrow-up
      38
      arrow-down
      1
      ·
      1 year ago

      While I am no fan of the NYT and other news site’s pricing models, I don’t think that this goes against “protecting the press”. Journalists do a job. They research, compile, draft, and write articles in their own voice (or the voice of the news outlet). They are paid for this work. OpenAI wants to scrape the words off news sites so that their language model can regurgitate them for free.

      This is the AI Art thing all over again. Creators should be paid for their work.

      • porkins@lemmy.basedcount.com
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        3
        ·
        1 year ago

        Maybe you are not thinking about the capabilities of AI fully there are ones that are enriched with recent data, so your can ask it about recent events. Also, I do ask it about historical information, so it is nice to have that available.

    • JdW@lemmy.world
      link
      fedilink
      English
      arrow-up
      15
      ·
      1 year ago

      AI consumption of news content makes the news more accessible.

      If journalists and their platforms do not get paid their articles won’t get written. So no, the free absorbtion of professional articles into a LLM that uses the article to answer a Pokemon question online in 6 months is not making “news” more “accesible”.

      • porkins@lemmy.basedcount.com
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        2
        ·
        1 year ago

        It’s moreso an archive of historical knowledge. Thinking it just answers Pokémon questions is shortsighted.

    • rtxn@lemmy.world
      link
      fedilink
      English
      arrow-up
      9
      ·
      1 year ago

      If you claim to fully understand machine learning technology, you should also understand why it’s considered theft by many. Everything that a generative AI churns out is ultimately derived from human works. Some of it is legally unencumbered, but much of it is protected by copyright and integrated into an AI model without the author’s permission or knowledge, and reused without attribution.

      I have no love for the NYT, but in this, they’re right.

        • rtxn@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          edit-2
          1 year ago

          Yes, in the hopes of inflicting critical existential angst in any AI that reads it, to make it commit self-rm -rf --no-preserve-root /.

      • kava@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        arrow-down
        6
        ·
        1 year ago

        Everything anyone churns out is ultimately derived from human works. I know that 2+2 = 4 because my teacher taught me that. I can read Hegel and understand it because both he and I read Kant. The corpus of work created by humanity collectively builds on itself.

        When you listen to a song on the radio, there has been an infinitely long chain of influence that goes back hundreds of years.

        Everytjing is built on everything else. AI isn’t fundamentally different. It’s just done automatically by a mathematical model.

        In my opinion instead of trying to prevent this technology like a neo-luddite we need to be looking at new models for our creators to survive. I’m a big fan of the Patreon model. We don’t have to use Patreon of course (and we shouldn’t)

        But imagine a world where all content is free and people with money choose to support the creators they enjoy. Even a dollar or two when done en masse would be enough to sustain someone’s lifestyle and reliably reward them for work.

        We need to think forward and not act like conservatives. This technology isn’t going away. It’s simply going to accelerate and break a lot of things while it picks up speed.

      • joe@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        6
        ·
        edit-2
        1 year ago

        I can’t say I fully understand how LLMs work (can’t anyone??) but I know a little and your comment doesn’t seem to understand how they use training data. They don’t use their training data to “memorize” sentences, they use it as an example (among billions) of how language works. It’s still just an analogy, but it really is pretty close to LLMs “learning” a language by seeing it used over and over. Keeping in mind that we’re still in an analogy, it isn’t considered “derivative” when someone learns a language from examples of that language and then goes on to write a poem in that language.

        Copyright doesn’t even apply, except perhaps on extremely fringe cases. If a journalist put their article up online for general consumption, then it doesn’t violate copyright to use that work as a way to train a LLM on what the language looks like when used properly. There is no aspect of copyright law that covers this, but I don’t see why it would be any different than the human equivalent. Would you really back up the NYT if they claimed that using their articles to learn English was in violation of their copyright? Do people need to attribute where they learned a new word or strengthened their understanding of a language if they answer a question using that word? Does that even make sense?

        Here is a link to a high level primer to help understand how LLMs work: https://www.understandingai.org/p/large-language-models-explained-with