• TheGrandNagus@lemmy.world
    link
    fedilink
    English
    arrow-up
    35
    arrow-down
    4
    ·
    edit-2
    2 days ago

    So people are complaining that the depictions of cardiologists, when you ask an image generator to show you one, are too accurate? That 85% of the time they show you a man in a job where 85% of people are men?

    You’d likely see the same thing if you asked one to show you a warehouse worker, another job that’s male-dominated.

    If people want more women in these roles, push for it at the university level. Push for it in medical posters in hospitals. I don’t see how forcing the hundreds of AI models out there to be biased in favour of depicting women when their training material doesn’t have as many is an effective way of achieving this goal.

    This just seems like a “we want to complain about this field being male dominated, and we’re sure to get headlines if we include the AI buzzword”

    • Deceptichum@quokk.au
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      20
      ·
      edit-2
      2 days ago

      Because the biases in an AI model will shape the perception of people who may think about entering those fields more than a poster at a place where people have already entered those fields work at.

      Likewise you can train it out of a bias, just feed it more content showing diverse workforces and it will start weighing them higher.

        • Deceptichum@quokk.au
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          15
          ·
          edit-2
          2 days ago

          People are not statistics.

          There are huge biases at play for why gender differences exist in the world place and most other places.

          Those biases carry on into the works we make, such as LLMs which are becoming hugely influential. To tackle these biases, we need to change how we view them. That means if the statistical average scientist is a white man, that we show more women and PoV to help them feel like they too can do this.

          tl;dr: We don’t need to reflect society as it currently is, we should aim to show how it can be.

          • conciselyverbose@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            14
            ·
            edit-2
            2 days ago

            Statistics are people.

            It cannot possibly, under any circumstances, be a correct, reasonable, or valid word choice to describe an “AI” that accurately models reality as “biased”. That word already has a meaning and using it in that manner is a lie.

            Bias is an irrational departure from reality. You can want an AI to be biased towards diversity, but that is adding bias, not removing it.

            • Randomgal@lemmy.ca
              link
              fedilink
              English
              arrow-up
              2
              ·
              1 day ago

              This. I’d you don’t like the AI presents more men than women as cardiologists because its mirroring society, the problem is not the AI, the problem is the lack of non cis-male representation in real world cardiology.

      • TheGrandNagus@lemmy.world
        link
        fedilink
        English
        arrow-up
        13
        ·
        edit-2
        2 days ago

        The depiction aligning with reality is not a bias. Artificially altering the algorithm so that it shows more women for this prompt on the other hand is, unquestionably, adding a bias.

        If you want to add a bias, fine. Biases aren’t always a bad thing, I can certainly see the argument for why you might want a 50/50 gender split for all AI prompts. But don’t pretend that what you’re actually advocating for here is correcting a bias, because it isn’t.

        Likewise you can train it out of a bias, just feed it more content showing diverse workforces

        That is training in a bias. Because it’s not representative of reality.

  • slacktoid@lemmy.ml
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 day ago

    AI is gonna mirror our biases back to us. If the AI has found a bias it’s cause or our internal biases. You could force diversity to hide the underlying issue and then you’ll get PoC nazis during WW2 like what Google did. Or … Call me crazy … But we could try to address the underlying societal issues.

  • Kusimulkku@lemm.ee
    link
    fedilink
    English
    arrow-up
    11
    ·
    1 day ago

    The gender distribution was not statistically different from that of actual Australian workforce data

    Sounds alright to me.