• Gaywallet (they/it)@beehaw.orgOP
      link
      fedilink
      arrow-up
      31
      ·
      edit-2
      11 months ago

      You’re absolutely correct, yet ask someone who’s very pro AI and they might dismiss such claims as “needing better prompts”. Also many people may not be as tech informed as you are, and bringing light to algorithmic bias can help them understand and navigate the world we now live in. Dismissing the article just because you already know the answer doesn’t really encourage people to participate in a discussion.

        • MBM@lemmings.world
          link
          fedilink
          arrow-up
          7
          ·
          11 months ago

          If you genuinely don’t know: because it’s an attention-grabbing title (which isn’t inherently bad)

      • Even_Adder@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        8
        ·
        11 months ago

        It’s really hard getting dark skin sometimes. A lot of the time it’s not even just the model, LoRAs and Textual Inversions make the skin lighter again so you have to try even harder. It’s going to take conscious effort from people to tune models that are inclusive. With the way media is biased right now, I feel like it’s going to take a lot of effort.

        • jarfil@beehaw.org
          link
          fedilink
          arrow-up
          3
          ·
          11 months ago

          “Inclusive models” would need to be larger.

          Right now people seem to prefer smaller quantized models, with whatever set of even smaller LoRAs on top, that make them output what they want… and only include more generic elements in the base model.

            • jarfil@beehaw.org
              link
              fedilink
              arrow-up
              2
              ·
              11 months ago

              Are you ready to run a 100B FP64 parameter model? Or even a 10B FP32 one?

              Over time, I wouldn’t be surprised if 500B INT8 models became commonplace with neuromorphic RAM, but there’s still some time for that to happen.

                • jarfil@beehaw.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  11 months ago

                  For more inclusive models, or for current ones? In order to add something, either the size has to grow, or something would need to get pushed out (content, or quality). 4GB models are already at the limit of usefulness, both DALLE3 and SDXL run at about 12B parameters, so to make them “more inclusive” they’d have to grow.

                  • Even_Adder@lemmy.dbzer0.com
                    link
                    fedilink
                    English
                    arrow-up
                    3
                    ·
                    11 months ago

                    I’m saying SD 1.5 and SDXL capture the concepts just fine, it’s just during fine-tuning people train away some of the diversity.

          • Muehe@lemmy.ml
            link
            fedilink
            arrow-up
            1
            ·
            11 months ago

            “Inclusive models” would need to be larger.

            [citation needed]

            To my understanding the problem is that the models reproduce biases in the training material, not model size. Alignment is currently a manual process after the initial unsupervised learning phase, often done by click-workers (Reinforcement Learning from Human Feedback, RLHF), and aimed at coaxing the model towards more “politically correct” outputs; But ultimately at that time the damage is already done since the bias is encoded in the model weights and will resurface in the outputs just randomly or if you “jailbreak” enough.

            In the context of the OP, if your training material has a high volume of sexualised depictions of Asian women the model will reproduce that in its outputs. Which is also the argument the article makes. So what you need for more inclusive models is essentially a de-biased training set designed with that specific purpose in mind.

            I’m glad to be corrected here, especially if you have any sources to look at.

            • jarfil@beehaw.org
              link
              fedilink
              arrow-up
              2
              ·
              11 months ago

              You can cite me on this:

              First, there is no thing as a “de-biased” training set, only sets with whatever target series of biases you define for them to reflect.

              Then, there are only two ways to change the biases of a training set:

              1. either you replace data until your desired objective, which will reduce the model’s quality for any of the alternatives
              2. or you add data until your desired objective, which will require an increased size to encode the increased amount of data, or the model’s quality will go down for all cases (you’d be diluting every other case)

              For reference, LoRAs are a sledgehammer approach to apply the first way.


              As for the article, it’s talking about the output of some app, with unknown extra prompting and LoRAs getting applied in the back, so it’s worthless as a discussion of the underlying model, much less as a discussion of all models.

              • Muehe@lemmy.ml
                link
                fedilink
                arrow-up
                1
                ·
                11 months ago

                First, there is no thing as a “de-biased” training set, only sets with whatever target series of biases you define for them to reflect.

                Yes, I obviously meant “de-biased” by definition of whoever makes the set. Didn’t think it worth mentioning, as it seems self evident. But again, in concrete terms regarding the OP this just means not having your dataset skewed towards sexualised depictions of certain groups.

                1. either you replace data until your desired objective, which will reduce the model’s quality for any of the alternatives

                […]
                For reference, LoRAs are a sledgehammer approach to apply the first way.

                The paper introducing LoRA seems to disagree (emphasis mine):

                We propose Low-Rank Adaptation, or LoRA, which freezes the pre-trained model weights and injects trainable rank decomposition matrices into each layer of the Transformer architecture, greatly reducing the number of trainable parameters for downstream tasks.

                There is no data replaced, the model is not changed at all. In fact if I’m not misunderstanding it adds an additional neural network on top of the pre-trained one, i.e. it’s adding data instead of replacing any. Fighting bias with bias if you will.

                And I think this is relevant to a discussion of all models, as reproduction of training set biases is something common to all neural networks.

                • jarfil@beehaw.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  edit-2
                  11 months ago

                  That paper is correct (emphasis mine):

                  We propose Low-Rank Adaptation, or LoRA, which freezes the pre-trained model weights and injects trainable rank decomposition matrices into each layer of the Transformer architecture, greatly reducing the number of trainable parameters for downstream tasks.

                  You can see how it works in the “Introduction” section, particularly figure 1, or in this nice writeup:

                  https://dataman-ai.medium.com/fine-tune-a-gpt-lora-e9b72ad4ad3

                  LoRA is a “space and time efficient” technique to produce a modification matrix for each layer. It doesn’t introduce new layers, or add data to any layer. To the contrary, it’s bludgeoning all the separate values in each layer, and modifying each whole column and whole row by the same delta (or only a few deltas, in any case with Ar«Wk and Br«Wd).

                  Turns out… that’s enough to apply some broad strokes type of changes to a model, which still limps along thanks to the remaining value variation. But don’t be mistaken: with each additional LoRA applied, a model loses some of its finer details, until at some point it descends into total nonsense.

                  • Muehe@lemmy.ml
                    link
                    fedilink
                    arrow-up
                    1
                    ·
                    11 months ago

                    Yeah but that’s my point, right?

                    That

                    1. you do not “replace data until your desired objective”.
                    2. the original model stays intact (the W in the picture you embedded).

                    Meaning that when you change or remove the LoRA (A and B), the same types of biases will just resurface from the original model (W). Hence “less biased” W being the preferable solution, where possible.

                    Don’t get me wrong, LoRAs seem quite interesting, they just don’t seem like a good general approach to fighting model bias.

    • Admetus@sopuli.xyz
      link
      fedilink
      arrow-up
      10
      ·
      11 months ago

      And every single Asian game and anime tends to go for skimpy or virtual softcore with it’s female characters. Rarely you see a female character in full armor.