• ArchRecord@lemm.ee
    link
    fedilink
    English
    arrow-up
    108
    arrow-down
    1
    ·
    3 months ago

    If OpenAI wants a pass, then just like how piracy services make content freely open and available, they should make their models open.

    Give me the weights, publish your datasets, slap on a permissive license.

    If you’re not willing to contribute back to society with what you used from it, then you shouldn’t exist within society until you do so.

    • CrayonMaster@midwest.social
      link
      fedilink
      English
      arrow-up
      57
      ·
      3 months ago

      Piracy steals from the rich and gives to the poor. ChatGPT steals from the rich and the poor and keeps for itself.

      • ArchRecord@lemm.ee
        link
        fedilink
        English
        arrow-up
        8
        ·
        3 months ago

        and keeps for itself.

        Which is why they should be legally compelled to publicize all of their datasets, models, research, and share any profits they’ve made with the works they can get provenance data for, because otherwise, it’s an unfair use of the public sphere of content.

        One could very easily argue that adblockers are piracy, and those would be stealing from every social media creator, small blog, and independent news site, but I don’t see many people arguing against that, even though that very well includes people who aren’t wealthy corporations.

        The issue isn’t necessarily the use of the copyrighted content, it’s the unfair legal stance taken on who can use the content, and how they are allowed to profit (or not profit) from it.

        I’m not saying there are no downsides, but I do feel like a simple black and white dichotomy doesn’t properly outline how piracy and generative AI training are relatively similar in terms of who they steal from, and it’s more of a matter of what is done with the content after it is taken that truly matters most.