• gmtom@lemmy.world
    link
    fedilink
    arrow-up
    4
    ·
    6 months ago

    Not sure if someone else has brought this up, but this is because these AI models are massively biased towards generating white people so as a lazy “fix” they randomly add race tags to your prompts to get more racially diverse results.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      6 months ago

      Exactly. I wish people had a better understanding of what’s going on technically.

      It’s not that the model itself has these biases. It’s that the instructions given them are heavy handed in trying to correct for an inversely skewed representation bias.

      So the models are literally instructed things like “if generating a person, add a modifier to evenly represent various backgrounds like Black, South Asian…”

      Here you can see that modifier being reflected back when the prompt is shared before the image.

      It’s like an ethnicity AdLibs the model is being instructed to fill out whenever generating people.

    • Marcbmann@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      6 months ago

      I mean, I don’t think it’s an easy thing to fix. How do you eliminate bias in the training data without eliminating a substantial percentage of your training data. Which would significantly hinder performance.

      • bamboo@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        0
        ·
        6 months ago

        Rather than eliminating the some of the training data, you could add more training data to create an even balance.

  • Pendulum@lemmy.world
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    6 months ago

    It’s horrifically bad, even if not compared against other LLMs. I asked it for photos of actress and model Elle Fanning (aged 25 or so) on a beach, and it accused me of seeking CSAM… That’s an instant never-going-to-use-again for me - mishandling that subject matter in any way is not a “whoopsie”

    My purpose is to help people, and that includes protecting children. Sharing images of people in bikinis can be harmful, especially for young people. I hope you understand.

  • Kusimulkku@lemm.ee
    link
    fedilink
    arrow-up
    0
    ·
    6 months ago

    This is fucking ridiculous. This AI is the worst of them all. I don’t mind it when they subtly try to insert some diversity where it makes sense but this is just nonsense.

    • Flumpkin@slrpnk.net
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      6 months ago

      They are experimenting and tuning. Apparently without any correction there is significant racist bias. Basically the AI reflects the long term racial bias in the training data. According to this BBC article it was an attempt to correct this bias but went a bit overboard.

      PS: I find it hilarious. If anything it elevates the AI system to art, since it now provides an emotionally provoking mirror about white identity.

      • Ottomateeverything@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        6 months ago

        Apparently without any correction there is significant racist bias.

        This doesn’t make it any less ridiculous. This is a central pillar of this kind of AI tech, and they’re trying to shove a band aid over the most obvious example of it. Clearly, that doesn’t work. It’s also only even attempting to fix one of the “problems” - they’re never going to be able to “band aid” every single place where the AI exhibits this problem, so it’s going to leave thousands of others un-fixed. Even if their band aid works, it only continues to mask the shortcomings of this tech and makes it less obvious to people that it’s horrendously inacurrate with the other things it does.

        Basically the AI reflects the long term racial bias in the training data. According to this BBC article it was an attempt to correct this bias but went a bit overboard.

        Exactly. This is a core failing of LLM tech. It’s just going to repeat all the shit it was fed to it. You’re never going to fix that. You can attempt to steer it in different directions, but the reason this tech was used was because it is otherwise impossible for us to trudge through all the info that was fed to it. This was the only way to get it to “understand” everything. But all of it’s understandings are going to have these biases, and it’s going to be just as impossible to run through and fix all of these. It’s like you didn’t have enough metal to build the titanic so you just built it out of Swiss cheese and are trying to duct tape one hole closed so it doesn’t sink. It’s just never going to work.

        This being pushed as some artificial INTELLIGENCE is the problem here. This shit doesn’t understand what it’s doing, it’s just regurgitating the things it’s consumed. It’s going to be exactly as flawed as whatever was put into it, and you can’t change that. The internet media it was trained on is racist, biased, full of undeniably false information, and massively swayed by propaganda on all sides of the fence. You can’t expect LLMs to do anything different when trained on that data. They’re going to have all the same problems. Asking these things to give you any information is like asking the average internet user what the answer is. And the average internet user is not very intelligent.

        These are just amped up chat bots with data being sourced from random bits of the internet. Calling them artificial INTELLIGENCE misleads people into thinking these bots are smart of have some sort of understanding of what they’re doing. They don’t. They’re just fucking internet parrots, and they don’t have the architecture to be “fixed” from having these problems. Trying to patch these problems out is a fools errand and only masks their underlying failings.

        • KeenFlame@feddit.nu
          link
          fedilink
          arrow-up
          0
          arrow-down
          1
          ·
          6 months ago

          None of this has been pushed, by any researcher, by any company, by any open source group even, as “intelligence” In fact, it was unanimously disliked as a term by everyone working with the models and transformers, but media circus combined with techbros laymen hard on hype have won. Since then everyone has given up trying to be semantically correct on this front.

          • Ottomateeverything@lemmy.world
            link
            fedilink
            arrow-up
            0
            ·
            6 months ago

            I didn’t say any researcher or anything had named it intelligence. Nor am I trying to be semantically correct.

            Read the guys comments. He’s trying to push the idea that we can “change” it’s “understanding” about the things it’s discussing. He is one of the people who has fallen for the tech bros etc convincing people it is intelligent. I’m not fighting semantics, I’m trying to explain to him that it’s not intelligent. Because he himself clearly doesn’t understand that.

            • KeenFlame@feddit.nu
              link
              fedilink
              arrow-up
              0
              arrow-down
              1
              ·
              6 months ago

              That’s just silly, as if there is no nuance whatsoever. You can ofc change its understanding. Depending on your definition, different types of models could be interpreted as intelligent in certain areas. You can be rational, you know, not everything needs to be black and white. It’s also possible that since even the experts in the field don’t fully grasp it, maybe you don’t either.

      • Kusimulkku@lemm.ee
        link
        fedilink
        arrow-up
        0
        ·
        edit-2
        6 months ago

        For example, a prompt seeking images of America’s founding fathers turned up women and people of colour.

        “A bit” overboard yeah

        • KeenFlame@feddit.nu
          link
          fedilink
          arrow-up
          0
          arrow-down
          1
          ·
          6 months ago

          To the machine, the query is “draw the founding fathers but diversely” it’s not the data that is corrupt, the usage is, clearly the system prompt in this case

  • Eddyzh@lemmy.world
    link
    fedilink
    arrow-up
    0
    ·
    6 months ago

    It is ridiculous. However, how can we know you did not first instruct to only show dark skin? Or select these from many examples that showed something else?

    • stoneparchment@possumpat.io
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      6 months ago

      It’s also like, I guess I would prefer it to make mistakes like this if it means it is less biased towards whiteness in other, less specific areas?

      Like, we know these models are dumb as rocks. We know that they are imperfect and that they mirror the biases of their trainers and training data, and that in American society that means bias towards whiteness. If the trainers are doing what they can to prevent that from happening, whatever, that’s cool… even if the result is some dumb stuff like this sometimes.

      I also don’t think it’s a problem for the user to specify race if it matters? Like “a white queen of England” is a fine thing to ask for, and if it isn’t specified, the model will include diverse options even if they aren’t historically accurate. No one gets bent out of shape if the outfits aren’t quite historically accurate, for example

      • ji59@kbin.social
        link
        fedilink
        arrow-up
        0
        ·
        6 months ago

        The problem is that these answers are hugely incorrect and if some child learning about history of England would see this, they would create bias that England was always diverse.
        The same is true for some recent post, where people knowing nothing about Scotland history could learn from images that half of Scotland population in 18th century was black.
        So from my perspective these images are just completely wrong and it should be fixed.
        Also if you want diversity, what about handicapped people?

        • groet@feddit.de
          link
          fedilink
          arrow-up
          0
          ·
          6 months ago

          Repeat after me:

          “Current AI is not a knowledge tool. It MUST NOT be used to get information about any topic!”

          If your child is learning Scottish history from AI, you failed as a teacher/parent. This isn’t even about bias, just about what an AI model is. It’s not even supposed to be correct, that’s not what it is for. It is for appearing as correct as the things it has been trained on. And as long as there are two opinions in the training data, the AI will gladly make up a third.

        • stoneparchment@possumpat.io
          link
          fedilink
          arrow-up
          0
          ·
          edit-2
          6 months ago
          • it’s true that this would mislead children, but the model could hallucinate about literally anything. Especially at this stage, no one-- children or adults-- should be uncritically accepting what the model states as fact. That said, I agree LLMs need to improve their factual accuracy

          • Although it is highly debated, some scholars suggest Queen Charlotte might have had African ancestry, or that she would be considered a POC by today’s standards. Of course, she reigned in the 17-1800s, but it isn’t entirely outlandish to have a “Queen of Color”, if we aren’t requesting a specific queen or a specific race

          • People of color did live in England in the middle ages? Like not diverse in the way we conceive now, but here are a few papers discussing the racial diversity at the time. It was surely less intermingled than today, but it’s not like these images are impossible

          • Other things are anachronistic or fantastical about these images, such as clothing. Are we worried about children getting the wrong impression of history in that sense?

          • Of course increasing visibility and representation of all kinds of marginalized people is important. I, myself, am disabled, so I care about that representation too-- thanks for pointing out how we could improve the model further. I do kinda feel like people would be groaning if the model had produced a Queen with a visible disability, though… I would be delighted to be wrong on this front :)

          • KeenFlame@feddit.nu
            link
            fedilink
            arrow-up
            0
            arrow-down
            1
            ·
            6 months ago

            These are not hallucinations. The image generator system prompt has been intensely altered to mix all races and genders. The model is probably not inaccurate up until it being misused. The misuse could be at any level of interaction, so it’s very misleading to base it on such an example

  • Amaltheamannen@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    6 months ago

    And how do we know you didn’t crop out an instruction asking for diversity?

    Either that or a side effect of trying to have less training data bias.

  • iain@feddit.nl
    link
    fedilink
    arrow-up
    0
    ·
    6 months ago

    This just shows that AI sucks for getting accurate information. Even if it didn’t hallucinate black people, it would’ve been just as wrong, just with white skinned queens. Now the lies just line up with “current social freakout of conservatives”.

    • KeenFlame@feddit.nu
      link
      fedilink
      arrow-up
      0
      arrow-down
      1
      ·
      6 months ago

      It really does not, even if you have a perfectly accurate model and ask it “draw an English queen, but make it ethnically diverse” this would still appear.

  • kandoh@reddthat.com
    link
    fedilink
    arrow-up
    0
    ·
    6 months ago

    Always telling when you see people online with a huge problem when AI generators aren’t racist or attempt to avoid racism.

    It’s almost like they see racism in technology as a sort of affirmation.