• JustARaccoon@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    3 months ago

    Are you done putting words in my mouth? Where did I say anything from the arguments you’re fighting against? I couldn’t give less of a shit what open ai wants, I’m not fighting for open ai, I’m fighting for all the artists who’ve been told again and again copyright infringement against big corpos is a no-no but now we have companies doing the same thing to them and it’s treated as an inevitability. For all I care open ai should be investigated for profiting from data they acquired through the loophole of being non-profit.

    What do any of the concerns over the way data acquisition happens have to do with open source? Open source the software, acquire the data ethically. Prosecute anyone using datasets with unlawfully acquired data to the same extent you’d prosecute copyright infringement because that’s what it is. No middle ground. There’s a shit ton of data in the public domain, use that instead of scouring artstation and written books from living writers. Is it not easy to sort or of less quality? Boohoo. If you want better data pay the artists and writers.

    Instead of this doomerposting “we’re gonna get the short stick either way might as well get something fun out of it” is exactly why we’re having our livelihoods trampled over.

    • Grimy@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      3 months ago

      I couldn’t give less of a shit what open ai wants, I’m not fighting for open ai, I’m fighting for all the artists

      What you want and what openai want are the same thing. Regulations directly benefit them by giving them and Google a easy peasy monopoly. Artists are never getting a dime out of any of this, all the data is already owned by websites and data brokers.

      open ai should be investigated for profiting from data they acquired through the loophole of being non-profit.

      This is patently false, there isn’t a loop hole. Almost all ml projects use public facing data, it’s accepted and completely legal since it’s highly transformative. What do you think translation software or Shazam uses? You probably already use AI multiple times a week. I’m guessing you didn’t get mad when all the translators lost their job a decade ago.

      What do any of the concerns over the way data acquisition happens have to do with open source?

      How can a company actually open source anything if the costs are so insanely high. It’s already above a million in compute power for a foundation model, how many open source projects do you expect if reddit or getty gets to tack on an other 60 million. Even worse, Microsoft and Google will absolutely pay a premium to keep it out of the hands of their competition. And no, there is simply not enough data in the public domain and most of it shit tbh.

      You are missing the forest for the tree and this is by design. There’s a reason you are bombarded every day by ai bad articles, it’s to keep you mad about it so you don’t actually think about what these regulations mean.

      • JustARaccoon@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        3 months ago

        Again, regulation doesn’t imply current giants get to still reap the rewards of that training data. Look at how GDPR affected data storage and acquisition retroactively. Assuming only one is possible is a false narrative.

        Public facing doesn’t mean open source. We’ve had this discussion before on GitHub accessible source code. Just because it’s available to peruse doesn’t mean one is allowed to process that image and create derivates based on its data. Weird thing to point out about translation, do you have any idea who I am or are you just regurgitating talking points? How do you know whether I was/am offended by translators being replaced or not?

        I’m confused about the open source bit, what costs? I feel like you’re not explaining a key connection in your argument. If the barrier to development overall is acquiring data ethically saying that is a stance against open source is misleading, as it’s against any kind of such development not just the open source kind. We have museums and library full of public domain works, it most definitely is enough, it’s just not as commercially appealing as modern works, so if given the choice of course companies will choose the path that gives them more rewards especially when we don’t punish them for copyright infringement when they do.

        You make it sound like LLMs are the best thing since sliced bread and should be pursued at all costs no matter how much it steps on the little guys in the process, but my question is why? We live in a world plagued by costs of living, atrocities, and other fixable things, sure this advanced text and image prediction stuff is a fun toy but will it actually improve the quality of life of people? Artists and writers already struggle more than your usual workers to get good pay for their time, this stuff might be sometimes touted as democratising art or something but it’s clearly not the main outcome from putting this kind of tool out in a world where capitalising on your skills is what gives people a roof over their heads. In such a world it’s only worsening peoples quality of life in exchange for a bit of fun and some performance improvements at work.

        And please don’t call me “mad”, don’t imply I’m clouded by emotions when I’m most surely providing clear statements. Throughout this I’ve been arguing against your points, but you’ve been arguing against a made up persona that you’ve attributed to me too. Go argue with those people and when you’re ready to engage me then argue against my points.

        • Grimy@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          arrow-down
          2
          ·
          3 months ago

          No regulations is going to force them to retroactively take their current models offline.

          Public facing doesn’t mean open source.

          Never said it was but public facing means you can scrape and use it for ml projects. This has already been decided in courts of law. You can’t use data with personal information or data which needs an account to access. Peruse kaggle for a bit, it’s all scraped datasets.

          do you have any idea who I am

          I literally don’t, I’m assuming you are part of the 99.999 % of population that didn’t get upset just like I assume you have arms and legs.

          Did you get upset about translators online when it happened?

          I’m also assuming you use AI on a weekly basis like practically everyone else else.

          You can give me a detailed biography and a list of every device, software and app you use, and I’ll stop assuming. Its fine if I’m wrong, point it out but it feels like I’m assuming correctly and instead of admitting it, you would rather get offended.

          the open source bit

          Paying 20x more than it currently costs to train a model will affect how many models are trained and given away for free.

          public domain works, it most definitely is enough

          Not enough to give a usable and competitive product. What’s the point of gimping open source so openai cam get all that profit. The jobs will still be lost regardless of if we can run these models on our computer or if a subscription service is the only option.

          Artists and writers already struggle more than your usual workers.

          I can empathize, I know it sucks. But regulations won’t change any of that. Deviant art will sell its dataset, the artists won’t be compensated and they will still have a hard time because these tools will still be available.

          And please don’t call me “mad”

          You commented under my post with a trite catch phrases. The tone of your comments aren’t very nice. I don’t know you, I’m going off of how you are saying it and it’s coming off as angry.