LINK (archive.ph)


AI may be a buzzword on Wall Street, but on the West Coast it’s at the center of Hollywood’s biggest labor dispute in more than 50 years. Among those warning about the technology’s potential to cause harm is British actor and author Stephen Fry, who told an audience at the CogX Festival in London on Thursday about his personal experience of having his identity digitally cloned without his permission.

“I’m a proud member of [actors’ union SAG-AFTRA], as you know we’ve been on strike for three months now. And one of the burning issues is AI,” he said.

Actors’ union SAG-AFTRA, which has around 160,000 members, went on strike last month over pay, working conditions, and concerns related to the use of AI in the film industry. It joined the Writers Guild of America—a union representing thousands of Hollywood writers—which went on strike in early May, marking the industry’s biggest shutdown in more than six decades.

A key sticking point for actors on strike is the possibility that studios could use AI to make digitally replicate their image without compensating them fairly for using their likeness.

Speaking at a news conference as the strike was announced, union president Fran Drescher said AI “poses an existential threat” to creative industries, and said actors needed protection from having “their identity and talent exploited without consent and pay.”

During his speech at CogX Festival on Thursday, Fry played a clip to the audience of an AI system mimicking his voice to narrate a historical documentary.

“I said not one word of that—it was a machine. Yes, it shocked me,” he said. “They used my reading of the seven volumes of the Harry Potter books, and from that dataset an AI of my voice was created and it made that new narration.”

Fry—who has appeared in movies including Gosford Park, V for Vendetta, and The Hitchhiker’s Guide to the Galaxy—is the narrator of the British Harry Potter audiobooks, while actor Jim Dale narrated the American version of the series.

“What you heard was not the result of a mash up, this is from a flexible artificial voice, where the words are modulated to fit the meaning of each sentence,” Fry told the audience at CogX Festival on Thursday.

“It could therefore have me read anything from a call to storm parliament to hard porn, all without my knowledge and without my permission. And this, what you just heard, was done without my knowledge. So I heard about this, I sent it to my agents on both sides of the Atlantic, and they went ballistic—they had no idea such a thing was possible.”

Fry added that when he discovered his voice was being used in projects without his consent, he saw it as just the beginning of an emerging threat to creative talent, warning his angry agents: “You ain’t seen nothing yet.” “This is audio,” he said he told them. “It won’t be long until full deepfake videos are just as convincing.”

As AI technology has advanced, doctored footage of celebrities and world leaders—known as deepfakes—has been circulating with increasing frequency, prompting warnings from experts about artificial intelligence risks. Fry warned on Thursday that those technologies only had further to go.

“We have to think about [AI] like the first automobile: impressive but not the finished article,” he said, noting that when cars were invented no one could have envisioned how widespread they are today.

“Tech is not a noun, it is a verb, it is always moving,” he said. “What we have now is not what will be. When it comes to AI models, what we have now will advance at a faster rate than any technology we have ever seen. One thing we can all agree on: it’s a f***ing weird time to be alive.”

Not the first

Fry isn’t the only famous actor to publicly vocalize their concerns about AI and its place in the film industry.

At a U.K. rally held in support of the SAG-AFTRA strike over the summer, Emmy-winning Succession star Brian Cox shared an anecdote about a friend in the industry who had been told “in no uncertain terms” that a studio would keep his image and do what they liked with it.

“That is a completely unacceptable position,” Cox said. “And that is the position that we should be really fighting against, because that is the worst aspect. The wages are one thing, but the worst aspect is the whole idea of AI and what AI can do to us.”

Oscar winner Matthew McConaughey told Salesforce CEO Marc Benioff during a panel event at this year’s Dreamforce conference that he had concerns about the rise of AI in Hollywood.

“We have a real chance, if we are irresponsible, of cannibalizing ourselves and creating this digital god that we’ll bow to, and we’ll all of a sudden become tools of this tool,” he said.

Meanwhile, Star Trek and Mission Impossible star Simon Pegg has called AI “worrying” for actors.

“We’re looking at being replaced in some ways,” he said at the rally in London in July. “We have to be compensated and we have to have some say in how [our image is] used. I don’t want to turn up in an advert for something I disagree with… I want to be able to hang on to my image, and voice, and know where it’s going.”

A spokesperson for the Alliance of Motion Picture and Television Producers (AMPTP), the entertainment industry’s official collective bargaining representative, was not available for comment when contacted by Fortune.

  • flossdaily@lemmy.world
    link
    fedilink
    arrow-up
    169
    arrow-down
    3
    ·
    1 年前

    The much, much, much more concerning aspect of voice cloning technology is that it will be used to scam people on a massive scale.

    Imagine you get a call at 4am from a loved one who tells you that they are in an emergency situation and had to borrow a phone to call you. The beg you to venmo some money to a stranger’s account so that they can get their car fixed/get a plane ticket/pay someone back for giving them a lift/etc.

    You recognize your loved one’s voice. They can respond to your questions (because chatbot AI). They know details about your life (because social media). It’s the middle of the night. You’re scared and not thinking clearly.

    This technology all exists TODAY. In 10 or 20 years it’ll be so terrifyingly sophisticated, even the most wary people will be vulnerable to it.

    • Steeve@lemmy.ca
      link
      fedilink
      English
      arrow-up
      64
      arrow-down
      2
      ·
      1 年前

      Easy solution, don’t have any loved ones. Checkmate scam artists

        • Kitty Jynx@lemmy.world
          link
          fedilink
          arrow-up
          5
          ·
          1 年前

          Someone tried to scam my grandpa with that. He told “me” to enjoy rotting in jail then called me up to ask how jail was.

        • Buddahriffic@lemmy.world
          link
          fedilink
          arrow-up
          4
          ·
          1 年前

          “It’s about time they caught you! Oh wait, they don’t want the reward money, do they? Ah fuck it, they can’t unarrest you, tell them to get fucked!”

      • Final Remix@lemmy.world
        link
        fedilink
        arrow-up
        7
        ·
        1 年前

        Or if you do, make sure none of 'em are dumb enough to rely on “cash apps” like venmo. Even Zelle, through our bank is suspicious as shit.

    • massive_bereavement@kbin.social
      link
      fedilink
      arrow-up
      50
      ·
      1 年前

      That’s why I do like Gilbert Gottfried and do two voices: one in public and one for friends and family.

      It gets confusing when we dine outside.

    • MossBear@lemmy.world
      link
      fedilink
      arrow-up
      19
      arrow-down
      1
      ·
      1 年前

      They can’t get me if I live in a hole. Not a nasty, dirty, wet hole, filled with the ends of worms and an oozy smell, nor yet a dry, bare, sandy hole with nothing in it to sit down on or to eat: but a hobbit-hole, and that means comfort.

    • RGB3x3@lemmy.world
      link
      fedilink
      arrow-up
      16
      arrow-down
      1
      ·
      1 年前

      The solution that EVERYBODY needs to learn for something like that is to hang up and call them back using the contact you have in your phone. They can afford 10 seconds while you do that if they’re calling you for money. And if it isn’t them calling for money, well sorry for waking you up Frank, but an AI was posing as you asking for cash.

      • agent_flounder@lemmy.one
        link
        fedilink
        English
        arrow-up
        11
        arrow-down
        1
        ·
        1 年前

        That and a family or per person verification word or protocol or something.

        “Clumsy…”

        “Draconiquist!”

        • Neato@kbin.social
          link
          fedilink
          arrow-up
          3
          ·
          1 年前

          “Oh my god, agent_flouder, I was just in a car accident and they need the bank info to process the co-insurance so I can get the organ transplant that is expiring in minutes! I’ve lost a lot of blood, have a concussion and have forgotten our code word. PLEASE don’t do this right now or this might be the last time you hear from me…”

          In the capitalist hellscape that is the US, that isn’t that far fetched and with emotions high, I doubt it’s unlikely. On the other hand, I can see a news article that reads, “Man lets daughter die by refusing hospital critical information needed for transplant.”

          • Rai@lemmy.dbzer0.com
            link
            fedilink
            arrow-up
            2
            ·
            1 年前

            Lawl they don’t do insurance shit for emergencies like needing an organ or blood immediately, they deal with that shit after the operation

            • Neato@kbin.social
              link
              fedilink
              arrow-up
              3
              ·
              1 年前

              Yeah but this is America. And do you think the average person knows that, will remember it when their loved one calls them crying, and will have the temerity to actually refuse when there’s a time constraint?

      • Neato@kbin.social
        link
        fedilink
        arrow-up
        4
        arrow-down
        1
        ·
        1 年前

        Unless they are calling from the hospital, police station, borrowed a cell phone after a car accident, etc.

    • PoliticalAgitator@lemm.ee
      link
      fedilink
      arrow-up
      11
      arrow-down
      5
      ·
      1 年前

      Scam them out of what?

      90% of the world will be unemployed and fighting for whatever scraps of food are grown between the constant flooding and fires we had 100 years to prevent but it wasn’t profitable to.

      We will have to learn to live entirely without factual information as every form of communication becomes hopelessly compromised by corporations, governments and billionaire extremists.

      You won’t even be able to trust the people you meet in meatspace. You think Fox News addicts are fucked up? Wait until every piece of entertainment is propaganda that’s been personalised just for you.

      And when it inevitably turns to war and you’re put in charge of the big red button, will you even care if the order to press it has been deep faked by a death cult?

      Unemployment and petty scams are small fry. With this technology, we can end the world.

      • Its_not_Dave@lemmy.world
        link
        fedilink
        arrow-up
        7
        arrow-down
        2
        ·
        1 年前

        In the referenced scenario they had to borrow a phone to call you.

        Presumably their phone is out of battery, broken, stolen, or they’re in another country without service.

        • FlexibleToast@lemmy.world
          link
          fedilink
          arrow-up
          14
          ·
          1 年前

          So that means the “real” person definitely won’t answer their phone right? That all is useful for trying to confirm someone is who they say they are.

          • XTornado@lemmy.ml
            link
            fedilink
            arrow-up
            7
            arrow-down
            2
            ·
            1 年前

            Yeah it has the flaw that at that hour the real person might not answer tough… (If they shutdown the phone or mute it or whatever). But yeah that is the common approach.

        • perviouslyiner@lemm.ee
          link
          fedilink
          English
          arrow-up
          4
          ·
          1 年前

          Deviant had this fascinating/awful video about this kind of situation: https://www.youtube.com/watch?v=6ihrGNGesfI

          Here is a (really) top infosec expert saying that when someone you know is in jail, you absolutely have to turn off all your call filtering and spam filtering, because who knows what shitty system the facility they’ve been moved to today is using to route calls.

        • Black_Gulaman@lemmy.dbzer0.com
          link
          fedilink
          arrow-up
          3
          ·
          1 年前

          Well since it’s a scam, it means that the phone is really not broken or out of commission.

          Not readily believing anything you hear from something that is unusual or out of the ordinary will save you in the future.

          Call just to check. If the phone is unreachable, call someone close to that person, a wife, a son /daughter. Just think. If they have a wife or kids, why call you in the first place.

          Sometimes it’s good to be not overly trusting.

    • lobut@lemmy.ca
      link
      fedilink
      arrow-up
      5
      ·
      edit-2
      1 年前

      I’ve had friends fall for email scams. We’re from Canada and I remember one being that another friend of our was stuck in Wales and needed a bank transfer. I knew it was a scam but a few of my friends were worried. I said that I live in the UK now and can take a train to Wales if you REALLY want. They were still panicking and saying they should do it in case. I’m like, you can’t be serious!

      So yeah, it can already be bad and with ChatGPT they can pass the Turing test. All of our friends will probably test each other on our memories. “Tell me the name of your ex-gf and which year and how you broke again?”

    • gelberhut@lemdro.id
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 年前

      This type of scam exists and, unfortunately, works without voice cloning and social media in East Europe for years.

    • Ataraxia@sh.itjust.works
      link
      fedilink
      arrow-up
      2
      ·
      1 年前

      Most people don’t know what their loved ones sound like on the phone. This is already a scam and you should never believe someone calling you like that. You can ask them something only you two know or just tell them to call the police and that you’ll meet them at the police department or hospital or whatever. Never give out credit card info ect over the phone. Nobody would ever do that in a legit situation.

  • themeatbridge@lemmy.world
    link
    fedilink
    arrow-up
    52
    arrow-down
    3
    ·
    1 年前

    I’m not going to lie, the ability to have Stephen Fry narrate my daily schedule every morning is one of the things I am most excited for. Like, I understand why he’s upset about a movie using his voice without his permission, but I didn’t expect to get his permission for my thing, either.

    • agent_flounder@lemmy.one
      link
      fedilink
      English
      arrow-up
      15
      arrow-down
      1
      ·
      1 年前

      I mean if he wants to make a fortune and also make people giddily happy… I’d totally pay for that. Or like Andy Serkis (whose narration of The Hobbit is a fucking masterpiece, by the way).

      Probably couldn’t afford it though.

      • loobkoob@kbin.social
        link
        fedilink
        arrow-up
        7
        ·
        1 年前

        Andy Serkis is such a good audiobook narrator. I loved his version of Terry Pratchett’s Small Gods!

      • Microw@lemm.ee
        link
        fedilink
        arrow-up
        5
        ·
        1 年前

        Yeah, this is where the union demands are heading: if you use the AI replica of an actor’s voice for a project, you’ll need to pay for a license.

    • Flying Squid@lemmy.world
      link
      fedilink
      arrow-up
      8
      ·
      1 年前

      I already have Siri be an English man so it kind of sounds like a butler. I would pay for it to sound like Stephen Fry. Or Michael Gough.

    • djmarcone@lemm.ee
      link
      fedilink
      arrow-up
      6
      ·
      1 年前

      The Fry version of the audio books are fantastic. I too would like Stephen Fry narrate my day.

    • Neato@kbin.social
      link
      fedilink
      arrow-up
      5
      arrow-down
      1
      ·
      1 年前

      Fry could just license a company to use AI for a narrow purpose in this instance. He could or have his people quality control check it and get royalties on sales. He could even record new bits of voice for stuff not already covered adequately. This would be a way for AI to benefit both parties. Instead AI companies are stealing prior work and copying it wholesale.

  • Fedizen@lemmy.world
    link
    fedilink
    arrow-up
    38
    arrow-down
    1
    ·
    1 年前

    There really needs to be a “right to identity” that companies can’t pretend to be you without express permission on a per instance basis and roll it into fraud protection/identity theft laws.

    • Codex@lemmy.world
      link
      fedilink
      arrow-up
      10
      ·
      1 年前

      Arguably, enumerating such rights in the first place mainly benefits the wealthy and corporations because once encapsulated, such “rights” can be bought and sold.

      An established actor like Fry has a lot of leverage, and the union may win the recognition of those rights, but then that’s just putting them on the negotiating table. What studio, trying to launch a film franchise, wouldn’t get the exclusive Digital Identity Use Rights, in-perpetuity, solely for use in that character? Sure, RDJ is free to go make other movies and control how his image is otherwise used, but Marvel-Fox-Disney gets to keep making Iron Man content (starting the AI-replicated likeness) for all time. And if they want Iron Man to sell Big Macs, too fucking bad, shouldn’t have sold your rights so cheap. Leverage he didn’t have when Marvel was rebooting his career.

      • Fedizen@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        1 年前

        “in perpetuity” rights to use somebody else’s identity seems like it should be illegal tbh

  • NutWrench@lemmy.world
    link
    fedilink
    arrow-up
    37
    arrow-down
    1
    ·
    1 年前

    Actors need to be able to trademark their image and audio likeness or corporations will puppeteer them for free. Forever.

    • SCB@lemmy.world
      link
      fedilink
      arrow-up
      15
      arrow-down
      3
      ·
      1 年前

      I honestly think this is the way forward. Trademark the likeness, and studios can use the likeness while you get an upfront fee and royalties.

      Everyone wins.

      • Margot Robbie@lemmy.world
        link
        fedilink
        arrow-up
        11
        arrow-down
        2
        ·
        1 年前

        Expansion of copyright that way could set a very dangerous precedent and have unintended consequences.

        For example, does that mean Trump, who was once part of SAG, can claim royalties every time SNL/late night talk show shows a clip of him saying dumb things? (I’m pretty sure one of these shows used a combination of impressionist/deepfake to mock him too, and you KNOW Trump is petty enough to actually do it)

        • SCB@lemmy.world
          link
          fedilink
          arrow-up
          13
          ·
          edit-2
          1 年前

          Clips being used in such a way would fall under Fair Use and Parody, no?

          You do raise a good point about knock-off effects, but things like “clip shows” and similar segments are generally protected by existing law.

          Again, not a bad point overall, and definitely thought-provoking, but I think this one specific example has an “out.”

          • flambonkscious@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            1 年前

            Great debate and I wish I could add something to the discussion, but I’m only on my first cup of tea and the caffeine doesn’t hit like coffee…

      • CrowAirbrush@lemm.ee
        link
        fedilink
        arrow-up
        3
        ·
        1 年前

        Can we all do that then, for all that encompasses us. So that all these internet business also have to stop selling our info since it’s ours.

    • ParsnipWitch@feddit.de
      link
      fedilink
      arrow-up
      18
      arrow-down
      2
      ·
      edit-2
      1 年前

      It won’t happen because complex models need beefier hardware. And I don’t mean “my gaming PC is real good!” beefy, but more like “only companies/institutes/rich people can afford and maintain this kind of hardware” beefy.

      The majority of people will be dependent on companies offering access to their models and we know how that goes. Even when people are trying to train open models, this will be a power only already privileged people can wield. Further increasing the gap between rich school kid and poor school kid, rich school and poor school, rich country and poor country.

      Because, even though the current AI is not the Sci-Fi “actual AI” some people seem to think it is, it can be quite the powerful tool and potentially change how (rich) people work.

      • time_lord@lemmy.world
        link
        fedilink
        arrow-up
        4
        arrow-down
        3
        ·
        1 年前

        “only companies/institutes/rich people can afford and maintain this kind of hardware” beefy.

        So go rent some space on AWS for the weekend, it maybe costs you $500 to have a voice model. It’s nowhere near as cost prohibitive as you think. Heck, iOS is going to have it on-device. Maybe not as good as this voice model, but free on consumer level hardware.

    • Margot Robbie@lemmy.world
      link
      fedilink
      arrow-up
      1
      arrow-down
      2
      ·
      1 年前

      Once the tech of generative AI is out of the box, you can’t really put it back in the box again.

      (Gotta sneak in my movie reference today. 😉)

    • uglyduckling81@lemmy.world
      link
      fedilink
      arrow-up
      5
      arrow-down
      7
      ·
      1 年前

      It’s the low hanging fruit. Voice acting is easy. You just sit and talk into a microphone. It was always going to be first cab of the rank to replace.

      Any job that requires subjective analysis and opinion will be harder to replace. But it will go eventually as well.

      One day they will match up AI with autonomous machines and then manual labour will be completely replaced as well

      I’ve no idea what’s going to happen to most people when that happens. A few rich people controlling every job in the world. There will probably have to be a shift to proper communism so the state owns all the robots and AI and the people just exist in what the state provides.

      It’s going to be bad except for the very few at the top.

      • drislands@lemmy.world
        link
        fedilink
        arrow-up
        6
        ·
        1 年前

        Voice acting is easy. You just sit and talk into a microphone

        There is more to voice acting than just talking into a mic. Can you honestly say you’ve never seen an animated show, or played a video game, and noticed that one character is particularly good? Or particularly bad?

        • uglyduckling81@lemmy.world
          link
          fedilink
          arrow-up
          1
          arrow-down
          2
          ·
          edit-2
          1 年前

          It’s easy to replace. Not necessarily easy to do yourself. Voice simulation is trivial compared to things like replicating a full person or doing complex tasks that aren’t the same every time. Or making subjective judgements.

          It’s why voice acting will be gone in the near future.

    • dynamojoe@lemmy.world
      link
      fedilink
      arrow-up
      10
      ·
      1 年前

      I’d say it’s not just his intelligence and presentation. There’s a high level of bravery there. You never know how someone is going to react when you tell them their beliefs are suspect (or outright bullshit) and just saying this opens him up to derision or worse from any rando with a soapbox to stand on.

  • AlexWIWA@lemmy.ml
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    1
    ·
    1 年前

    I will never pay for an AI audio book. Never. Why would I pay a book company when I can train my own AI voice clone?

    Book publishers don’t realize how hard they’ll fuck themselves over if people realize how easy this is.

  • Tigbitties@kbin.social
    link
    fedilink
    arrow-up
    19
    ·
    1 年前

    Voice actor here. There’s an ad on the radio that sent in an audition tape for. I keep hearing it and I’m pretty sure it’s an AI voice. Thankfully it’s awful. Unfortunately, only a few people will notice or even care.

    • Swedneck@discuss.tchncs.de
      link
      fedilink
      arrow-up
      7
      ·
      1 年前

      It astounds me that people tolerate bad computer voices, i’m rather picky which real actual human voices i can tolerate listening to…

      • MisterD@lemmy.ca
        link
        fedilink
        arrow-up
        7
        ·
        1 年前

        I hear AI voices on YT videos all the f’n time. It’s monotonous, super consistent pauses, and not obvious until you listen for 30 secs or so.

    • Historical_General@lemm.eeOPM
      link
      fedilink
      arrow-up
      2
      ·
      1 年前

      Well, that’s depressing - to have it on the radio is an intrusion in real life in a way - outside the digital space.

  • qooqie@lemmy.world
    link
    fedilink
    arrow-up
    19
    arrow-down
    1
    ·
    1 年前

    Oh hell yeah it’s just the beginning. Go check out some popular twitch streamers, they have TTS of tons of celebs using their voices. Those idiots are gonna get sued so fucking fast once celebrities figure out computers

    • SupraMario@lemmy.world
      link
      fedilink
      arrow-up
      16
      arrow-down
      1
      ·
      1 年前

      So never then?

      Really hard to sue someone if you’re just using sound bites, I doubt actors give zero shits about twitch streamers.

      • qooqie@lemmy.world
        link
        fedilink
        arrow-up
        9
        ·
        1 年前

        Oh I definitely think they’ll care. It’s using their likeness for their own personal gain, the actors will want some of that money

        • SupraMario@lemmy.world
          link
          fedilink
          arrow-up
          4
          ·
          1 年前

          Unless some streamer is making millions using a single actors likeness. I don’t think thats who Fry is talking about here

          • qooqie@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            1 年前

            Some of these streamers are making millions off the donations they receive to use TTS on their stream

        • Riskable@programming.dev
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          1 年前

          You can’t copyright or trademark a voice. It’s 100% legal to clone someone’s voice and use it however your like. Basically, a person’s voice is public domain (according to the law as written right now).

          You may be thinking, “well, they’re just going to change the laws now that rich and powerful people are concerned about it.” It’s certainly possible that they’ll try to do that but in reality it’s not workable. Here’s why: What makes your voice your voice?

          A person’s voice (and accent) is unique for sure but using technological means to detect this is nearly impossible because there’s only so many possible variations of the human voice. Chances are there’s at least 80 million people with the exact same vocal chord sound as you. They all probably won’t have the same accent but that’s incredibly opinionated.

          If a law gets passed giving people the right to copyright or trademark their voice (i.e. the algorithmic signature of it) there will instantly be millions of completely Innocent “violators” who are suddenly having their videos and music removed from everything because their voice matches the signature. It’s not workable. At least not technologically.

          Another problem is that people’s voices change over time. Imagine if a kid’s voice was copyrighted… how many new children would suddenly be caught in that law enforcement net just by speaking? The voices of children are even less unique than adults!

      • Magrath@lemmy.ca
        link
        fedilink
        arrow-up
        3
        ·
        1 年前

        They aren’t sound bites of anything the celebrities have said. They are original clips generated from text submitted by the viewer who has paid a fee for that to happen. Some celebrities actually sell their services this way and the streamer is taking that revenue from them.

      • Kogasa@programming.dev
        link
        fedilink
        arrow-up
        1
        ·
        edit-2
        1 年前

        Well, you probably can copyright aspects of your likeness that are sufficiently distinct. Just your voice might be too generic, but your voice and accent and mannerisms etc. presented in context (e.g. Trump-bot used in a political ad) might be specific enough to be convincing in court. I don’t think streamers are getting sued (successfully) at any point though.

        I think/hope we eventually set some sane legal ground rules for the protection of individuals’ likenesses, causing (or stemming from) some high profile court case(s) involving serious offenders (e.g. publishing an AI-narrated audiobook without the original VA’s consent). Once that’s in the media for a bit (again I think/hope) we see companies like Twitch and Youtube ban the use of AI imitations on their platform to protect themselves. Just like DMCA.

        • You can not copyright aspects of your likeness. There are specific things that can be copyrighted in the US, and that is not one of them.

          There are laws against having your likeness used for commercial purposes but that gets pretty murky when it comes to voices.

      • ares35@kbin.social
        link
        fedilink
        arrow-up
        6
        ·
        1 年前

        “fair use” varies by jurisdiction and usually has a bit more than just sample size or length as a criteria.

        • TigrisMorte@kbin.social
          link
          fedilink
          arrow-up
          2
          arrow-down
          2
          ·
          1 年前

          Not really. You get the occasional loon that thinks they can copyright it unless it is a critique, but Courts have ruled the use transformative when it is a sound byte in a performance and not like a soundboard application or other such. And if it is a soundboard clip then the rights would be owed and paid by the company making the soundbyte and not the performer. The folks on twitch or whatever are on fairly safe ground.

          • Magrath@lemmy.ca
            link
            fedilink
            arrow-up
            1
            ·
            1 年前

            What about celebrities who sell personalized messages? They have a right to make money off their likeness, not anyone else unless they’ve been given permission. These streamers are making money off their likeness with no cut going to the celebrity.

            • TigrisMorte@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              1 年前

              Not remotely the same thing. A streamer playing a clip of the original Artist saying the famous line isn’t going to cause said Artist to want to attack the use. Playing Mark Hamill yelling, “You are not my Father!”, isn’t a personalized message nor is that what the streamer is selling.

      • Gigan@lemmy.world
        link
        fedilink
        arrow-up
        5
        ·
        1 年前

        Fair use is a legal defense, not a law. It doesn’t mean you can’t be sued and it doesn’t guarentee you’ll win if you are.

        • TigrisMorte@kbin.social
          link
          fedilink
          arrow-up
          1
          ·
          1 年前

          You need a Lawyer willing to take the case and a Judge that won’t laugh while sanctioning them. I did production work for Years and this is how their Lawyers explained it to me.

      • qooqie@lemmy.world
        link
        fedilink
        arrow-up
        1
        arrow-down
        1
        ·
        1 年前

        Oh shit I didn’t know that. Do you think that’ll apply to AI generated TTS?

  • Cosmic Cleric@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    1
    ·
    1 年前

    What we’re talking about is not so much AI itself but who owns the data set that the AI is created and trained with?

    Individuals should own the right to themselves, but if they want to sell it as a data set, then so be it.

    They can have restrictions as to how the AI that uses their data set to be created can be used and for what purposes.

    • Pika@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      8
      ·
      edit-2
      1 年前

      I disagree with this because companies will start enforcing a “if you wish to work for us you must give the ability to use your data set including after termination and etc with no further compensation”. This needs to be strictly a per individual per instance basis preferably requiring the person who owns it be the primary sale person included in the transaction.

      • Cosmic Cleric@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        1 年前

        Yeah it’ll definately take legislature to prevent scenarios like you described, but at the end of the day, a person should own their own identity, and nobody else should be able to make a 100% copy of it.

  • SmoothLiquidation@lemmy.world
    link
    fedilink
    arrow-up
    12
    arrow-down
    5
    ·
    1 年前

    There are going to be some laws hastily passed for this that is going to put impressionists out of the job. If it is Rich Little or and AI impersonating Howard Cosell, how is it any different?

    • flerp@lemm.ee
      link
      fedilink
      arrow-up
      5
      arrow-down
      2
      ·
      1 年前

      Impersonators are humans who also need to eat. Impersonating requires practice and talent. Impersonating doesn’t put the impersonated out of a job.

  • sebinspace@lemmy.world
    link
    fedilink
    arrow-up
    7
    arrow-down
    1
    ·
    1 年前

    Be real, Stephen. You’re voice wasn’t stolen from the audiobooks.

    It was stolen from every recording of your voice :D

  • emergencyfood@sh.itjust.works
    link
    fedilink
    arrow-up
    9
    arrow-down
    4
    ·
    1 年前

    Fry … is the narrator of the British Harry Potter audiobooks.

    Well, there’s your problem. Should have got Hatsune Miku herself to narrate the books she wrote.