Which of the following sounds more reasonable?

  • I shouldn’t have to pay for the content that I use to tune my LLM model and algorithm.

  • We shouldn’t have to pay for the content we use to train and teach an AI.

By calling it AI, the corporations are able to advocate for a position that’s blatantly pro corporate and anti writer/artist, and trick people into supporting it under the guise of a technological development.

  • aezart@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    arrow-down
    4
    ·
    1 year ago

    If an LLM was trained on a single page of GPL code or a single piece of CC-BY art, the entire set of model weights and any outputs from the model must be licensed the same way. Otherwise this whole thing is just blatant license laundering.

    • paperbenni@lemmy.world
      link
      fedilink
      English
      arrow-up
      13
      arrow-down
      2
      ·
      1 year ago

      This depends on how transformative the act of encoding the data in an LLM is. If you have overfitting out the ass and the model can recite its training material verbatim then it’s an illegal copy of the training material. If the model can only output content that would be considered transformative if a human with knowledge of the training data created it, then so is the model.

  • Iceblade@lemmy.world
    link
    fedilink
    English
    arrow-up
    12
    arrow-down
    1
    ·
    1 year ago

    IMO content created by either AI or LLMs should have a special license and be considered AI public domain (unless they can prove that they own all content the AI was trained on). Commercial content made based on content marked with this license would be subject to a flat % tax that should be applied to the product price which would be earmarked for a fund distributing to human creators (coders, writers, musicians etc.).

    • Trainguyrom@reddthat.com
      link
      fedilink
      English
      arrow-up
      7
      ·
      1 year ago

      I think the cleaner (and most likely) outcome is AI generated work is considered public domain, and since public domain content can already be edited and combined and arranged to create copyrighted content this would largely clear up the path for creators to use AI more prominently in their workflows

      • makingStuffForFun@lemmy.ml
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        1 year ago

        So I can make derivative works from commercial works, make something from that material, then release the result as public domain? I would think not.

      • Iceblade@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Honestly, I’d personally prefer the latter, but there is the argument made by artists, coders and content creators. Their work is being scraped to train these AI’s, which in turn makes their future work less valuable. Hence, the thought of enforcing a tiny “royalty”/tax on commercial products based off of AI generated content and funneling that money back to human creators of intellectual works.

    • kklusz@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      arrow-down
      2
      ·
      1 year ago

      What about LLM generated content that was then edited by a human? Surely authors shouldn’t lose copyright over an entire book just because they enlisted the help of LLMs for the first draft.

      • Cethin@lemmy.zip
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        2
        ·
        1 year ago

        If you take open source code using GNU GPL and modify it, it retains the GNU GPL license. It’s like saying it’s fine to take a book and just change some words and it’s totally not plagerism.

        • kklusz@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          Public domain is not infectious like GPL is. That being said, it seems like the parent comment has already mentioned this case, now that I’ve read them again:

          public domain content can already be edited and combined and arranged to create copyrighted content

          That’s fine by me. The important thing is that humans can still use AI as a legally recognized productivity tool, including using it as a way to use ideas and styles generated by other humans.

  • Fylkir@lemmy.sdf.org
    link
    fedilink
    English
    arrow-up
    15
    arrow-down
    4
    ·
    1 year ago

    I see it like this:

    Our legal system has the concept of mechanical licensing. If your song exists, someone can demand the right to cover it and the law will favor them. The result of an LLM has less to do with your art that a cover of your song does.

    There are plenty of cases of a cover eclipsing the original version of a song in popularity and yet I have never met a single person argue that we should get rid of the right to cover a song.

    • nosycat@forum.fail
      link
      fedilink
      arrow-up
      15
      arrow-down
      4
      ·
      1 year ago

      Sure, you have the legal right to cover someone else’s song without asking permission first, but you still have to pay them royalties afterwards, at fair market rates.

  • pensivepangolin@lemmy.world
    link
    fedilink
    English
    arrow-up
    9
    ·
    1 year ago

    I think it’s the same reason the CEO’s of these corporations are clamoring about their own products being doomsday devices: it gives them massive power over crafting regulatory policy, thus letting them make sure it’s favorable to their business interests.

    Even more frustrating when you realize, and feel free to correct me if I’m wrong, these new “AI” programs and LLMs aren’t really novel in terms of theoretical approach: the real revolution is the amount of computing power and data to throw at them.

    • assassin_aragorn@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      8
      ·
      1 year ago

      The funniest thing I’ve seen on this is the ChatGPT CEO, Altman, talking about how he’s a bit afraid of what they’ve created and how it needs limitations – and then when the EU begins to look at regulations, he immediately rejects the concept, to the point of threatening to leave the European market. It’s incredibly transparent what they’re doing.

      Unfortunately I don’t know enough about the technology to say if the algorithms and concepts themselves are novel, but without a doubt they couldn’t exist without modern computing power capabilities.

      • FancyGUI@lemmy.fancywhale.ca
        link
        fedilink
        English
        arrow-up
        3
        ·
        edit-2
        1 year ago

        I can tell for a fact that there’s nothing new going on. Only the MASSIVE investment from Microsoft to allow them to train on an insane amount of data. I am no “expert” per se, but I’ve been studying and working with AI for over a decade - so feel free to judge my reply as you please

        • Trainguyrom@reddthat.com
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          4
          ·
          1 year ago

          nothing new going on

          Uhhhh the available models are improving by leaps and bounds by the month, and there’s quite a bit of tangible advancement happening every week. Even more critically the models that can be run on a single computer are very quickly catching up to those that just a year or two ago required some percentage of a hyperscaler’s datacenter to operate

          Unless you mean to say that the current insane pace of advancement is all built off of decades of research and a lot of the specific advancements recently happen to be fairly small innovations into previous research infused with a crapload of cash and hype (far more than most researchers could only dream of)

          • FancyGUI@lemmy.fancywhale.ca
            link
            fedilink
            English
            arrow-up
            9
            ·
            edit-2
            1 year ago

            all built off of decades of research and a lot of the specific advancements recently happen to be fairly small innovations into previous research infused with a crapload of cash and hype>

            That’s exactly what I mean! The research projects I’ve been 5-7 years ago had already created LLMs like this that were as impressive as GPT. I don’t mean that the things that are going on aren’t impressive, I just mean that there’s nothing actually new. That’s all. IT’s similar to the previous hype wave that happened in AI with machine learning models when google was pushing deep learning. I really just want to point that out.

            EDIT: Typo

        • SCB@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          arrow-down
          5
          ·
          1 year ago

          nothing new going on

          I can’t think of anything less accurate to say about LLMs other than that they’re a world-ending threat.

          This is a bit like saying “The internet is a cute thing for tech nerds but will never go mainstream” in like 1995.

      • Peruvian_Skies@kbin.social
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        The concepts themselves are some 30 years old, but storage capacity and processing speed have only recently reached a point where generative AI outperforms competing solutions.

        But regarding the regulation thing, I don’t know what was said or proposed, and this is just me playing devil’s advocate: but could it be that the CEO simply doesn’t agree with the specifics of the proposed regulations while still believing that some other, different kind of regulation should exist?

        • rainh@kbin.social
          link
          fedilink
          arrow-up
          5
          ·
          1 year ago

          Certainly could be, but probably an optimistic take. Most likely they’re just trying to do what corporations have been doing for ages, which is to weaponize government policy to prevent competition. They don’t want restrictions that will materially impact their product, they want restrictions that will materially impact startups to make it more difficult for them to intrude on the established space.

          • jumperalex@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            ·
            1 year ago

            I think if you fed your response into ChatGPT and asked it to summarize in two words it would return,

            “Regulatory Capture”

    • ywein@lemmy.ml
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 year ago

      LLMs are pretty novel. They are made possible by invention of the Transformer model, that operates significantly different compared to, say, RNN.

    • eerongal@ttrpg.network
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      Even more frustrating when you realize, and feel free to correct me if I’m wrong, these new “AI” programs and LLMs aren’t really novel in terms of theoretical approach: the real revolution is the amount of computing power and data to throw at them.

      This is 100% true. LLMs, neural networks, markov chains, gradient descent, etc. etc. on down the line is nothing particularly new. They’ve collectively been studied academically for 30+ years. It’s only recently that we’ve been able to throw huge amounts of data, computing capacity, and time to tweak said models to achieve results unthinkable 10-ish years ago.

      There have been efficiencies, breakthroughs, tweaks, and changes over this time too, but that’s just to be expected. But largely its just sheer raw size/scale that’s just been achievable recently.

      • pensivepangolin@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Okay, I’m glad I’m not too far off the mark then (I’m not an AI expert/it’s not my field of study).

        I think this also points to/is a great example of another worrying trend: the consolidation of computing power in the hands of a few large companies. Without even factoring in the development of true AI/whether that can or will happen anytime soon, the LLMs really show off the massive scale of both computational power consolidation AMD data harvesting by only a very few entities. I’m guessing I’m not alone here in finding that increasingly concerning, particularly since a lot of development is driving towards surveillance applications.

      • Muehe@lemmy.ml
        link
        fedilink
        English
        arrow-up
        0
        ·
        1 year ago

        LLMs aren’t really novel in terms of theoretical approach: the real revolution is the amount of computing power and data to throw at them.

        This is 100% true. LLMs, neural networks, markov chains, gradient descent, etc. etc. on down the line is nothing particularly new. They’ve collectively been studied academically for 30+ years.

        Well LLMs and particularly GPT and its competitors rely on Transformers, which is a relatively recent theoretical development in the machine learning field. Of course it’s based in prior research, and maybe there even is prior art buried in some obscure paper or 404 link, but if that’s your measure then there is no “novel theoretical approach” for anything, ever.

        I mean I’ll grant that the available input data and compute for machine learning has increased exponentially, and that’s certainly an obvious factor in the improved output quality. But that’s not all there is to the current “AI” summer, general scientific progress played a non-minor part as well.

        In summary, I disagree on data/compute scale being the deciding factor here, it’s deep learning architecture IMHO. The former didn’t change that much over the last half decade, the latter did.

        • pensivepangolin@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 year ago

          Now as I stated in my first comment in these threads, I don’t know terribly much about the technical details behind current LLM’s and I’m basing my comments on my layman’s reading.

          Could you elaborate on what you mean about the development of of deep learning architecture in recent years? I’m curious; I’m not trying to be argumentative.

          • Muehe@lemmy.ml
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 year ago

            Could you elaborate on what you mean about the development of deep learning architecture in recent years?

            Transformers. Fun fact, the T in GPT and BERT stands for “transformer”. They are a neural network architecture that was first proposed in 2017 (or 2014 depending on how you want to measure). Their key novelty is the method of implementing an attention mechanism and a context window without recursion, which was the method most earlier NNs used for that.

            The wiki page I linked above is admittedly a bit technical, this articles explanation might be a bit more friendly to the layperson.

            • pensivepangolin@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              1 year ago

              Thanks for the reading material: I’m really not familiar with Transformers other than the most basic info. I’ll give it a read when I get a break from work.

      • jumperalex@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        1 year ago

        by that logic there was nothing novel about solid state transistors since they just did the same thing as vacuum tubes; no innovation there I guess. No new ideas came from finally having a way to pack cooler, less power hungry, smaller components together.

          • dismalnow@kbin.social
            link
            fedilink
            arrow-up
            1
            ·
            1 year ago

            AOL chatbot that did basic stuff.

            Here you go.

            If you ask Siri, “Do you sleep?” Siri will respond, “I don’t need much sleep, but it’s nice of you to ask.” Meanwhile, if you asked SmarterChild the same question, he would respond, “No, but I dream. I dream of a better world. A world where man and machine can coexist in peace and happiness.”

    • assassinatedbyCIA@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      It also plays into the hype cycle they’re trying to create. Saying you’ve made an AI is more likely to capture the attention of the masses then saying you have a LLM. Ditto that point for the existential doomerism that they ceo’s have. Saying your tech is so powerful that it might lead to humanity’s extinction does wonders in building hype.

      • pensivepangolin@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Agreed. And all you really need to do is browse any of the headlines from even respectable news outlets to see how well it’s working. It’s just article after article uncritically parroting whatever claims these CEO’s make at face value at least 50% of the time. It’s mind-numbing.

    • Phantom_Engineer@lemmy.ml
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      The fear mongering is pretty ridiculous.

      “AI could DESTROY HUMANITY. It’s like the ATOMIC BOMB! Look at it’s RAW POWER!”

      AI generates an image of cats playing canasta.

      “By God…”

    • RossoErcole@kbin.social
      link
      fedilink
      arrow-up
      0
      arrow-down
      1
      ·
      1 year ago

      We could say that the human brain isn’t novel in terms of biological composition: the real evolution is the size increase compared to the body.

      The fact that insects exist doesn’t make us less intelligent.

      But I agree with the sentiment of the argument.

  • Chocrates@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    1 year ago

    both sound the same to me IMO. Private companies scraping ostensibly public data to sell it. No matter how you word it they are trying to monetize stuff that is out in the open.

    • Dran@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      1 year ago

      I don’t see why a single human should be able to profit off learning from others but a group of humans doing it for a company cannot. This is just how humanity advances at whatever scale.

      • Chocrates@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        I had a comment about the morality of it at first but I pulled it out. This is not an easy question to answer. Corporations gate keeping knowledge seems weird and dystopian but the knowledge is out there and they are just making connections between it. It also touches on copyright and fair use.

        • Dran@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          I agree it’s much more complicated an issue than most people give it credit.

  • eerongal@ttrpg.network
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    1
    ·
    1 year ago

    I’m not sure what you’re trying to say here; LLMs are absolutely under the umbrella of AI, they are 100% a form of AI. They are not AGI/STRONG AI, but they are absolutely a form of AI. There’s no “reframing” necessary.

    No matter how you frame it, though, there’s always going to be a battle between the entities that want to use a large amount of data for profit (corporations) and the people who produce said content.

    • Silinde@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      True, and this is the annoying thing about people unqualified to talk about AI giving their opinions online. People not involved in the industry hear “AI” and expect HAL-9000 or Ava from Ex Machina rather than the software that the weather service uses to predict if it will rain tomorrow, or the models your doctor uses to help determine your risk of Heart Disease.

      This is compounded further when someone makes a video simplifying what an LLM is and mentioning that the latest models use it, which leads to the chimes of “bUt iT’S jUsT aN Llm BrO iTs nOt AI” and “ItS jUsT a LOaD oF DaTa aND aLGorItHMs, tHaTs NoT AI”. A little bit of knowledge is a dangerous thing.

      • assassin_aragorn@lemmy.worldOP
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        1 year ago

        This is actually exactly what I mean. Most people hear AI and envision something much, much more complex. It’s easier to argue that HAL-9000 is like a human and should therefore be allowed to freely view book content like a human, versus argue that a sophisticated LLM is like a human and should be allowed to freely view books like a human. That’s moreso where I’m coming from. And politicians are stupid enough to pass laws envisioning these as HAL-9000.

    • assassin_aragorn@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Right, where I’m coming from is that I don’t think the personhood arguments you see for why content should be free for it really hold any water. Whatever the case on its intelligence, it isn’t comparable to humans for copyright law

    • Peruvian_Skies@kbin.social
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      On the flip side, the same battle is also fought between giant corporations that amass intellectual property and the people who want to actually use that intellectual property instead of letting it sit in some patent troll’s hoard until a lawsuit op presents itself. Seeing as there are quite a few reasonably decent open-source LLMs out there like Koala and Alpaca also training on data freely available on the Internet, I’m actually rooting for the AI companies in this case, in the hopes of establishing a disruptive precedent.

  • Zeth0s@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    1 year ago

    That’s absolutely not correct. AI is a field of computer science/scientific computing built on the idea that some capabilities of biological intelligences could be simulated or even reproduced “in silicon”, i.e. by using computers.

    Nowadays is an extremely broad term that covers a lot of computational methodologies. LLM in particular are a evolution of methods born to simulate and act as human neural network. Nowadays they work very differently, but they still provide great insights on how an “artificial” intellicenge can be built. It is only one small corner of what will be a real general artificial intelligence, and a small step in that direction.

    AI as a name is absolutely unrelated with how programs based on the methodologies are built.

    Human intelligences are in charge of all copyright part. AI and copyright are orthogonal, people are those who cannot tell the 2 and keep talking about AI.

    There is AI, and there is copyright, it is time for all of us to properly frame the discussion on “copyright discussion related to <company>'s product”

    • assassin_aragorn@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      What I’m getting at moreso is that comparisons to humans for purposes of copyright law (e.g. likening it to students learning in school or reading library books) don’t hold water just because it’s called an AI. I don’t see that as an actual defence for these companies, and it seems to be somewhat prevalent.

      • Zeth0s@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        edit-2
        1 year ago

        You can absolutely compare AI with students. The problem is that, legally, in many western countries students still have to pay copyright holders of the books they use to learn.

        It is purely a copyright discussion. How far copyright applies? Shall the law distinguish between human learning and machine learning? Can we retroactively change copyright of material available online?

        For instance, copilot is more at risk than a LLM that learned from 4chan, because licenses are clearer there. Problem is that we have no idea on which data big llms were trained, to know if some copyright law already applies.

        At the end it is just a legal dispute on companies making money out of AI trained on data publicly available (but not necessarily copyright free).

        • assassin_aragorn@lemmy.worldOP
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          My argument is that an LLM here is reading the content for different reasons than a student would. The LLM uses it to generate text and answer user queries, for cash. The student uses it to learn their field of study, and then apply it to make money. The difference is that the student internalizes the concepts, while the LLM internalizes the text. If you used a different book that covered the same content, the LLM would generate different output, but the student would learn the same thing.

          I know it’s splitting hairs, but I think it’s an important point to consider.

          My take is that an LLM algorithm can’t freely consume any copyrighted work, even if it’s been reproduced online with the consent of the author. The company would need the permission of the author for the express purpose of training the AI. If there’s a copyright, it should apply.

          You have me thinking though about the student comparison. College students pay to attend lectures on material that can be found online or in their textbooks. Wouldn’t paying for any copyright material be analogous to this?

          • Zeth0s@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            arrow-down
            1
            ·
            edit-2
            1 year ago

            Students and LLM do the same with data, simply in a different way. LLM can learn more data, student can understand more concepts, logic and context.

            And students study to make money.

            Both LLMs and students map the data in some internal representation, that is however pretty different, because a biological mind is different from an AI.

            Regarding your last paragraph, this is exactly the point. What shall openai and Microsoft pay, as they are making a lot of money out of other people work? Currently it is unclear as openai hasn’t released what data they used, and because copyright laws do not cover generative AI. We need to wait for interpretation of existing laws and for new ones. But it will change soon in the future for sure

  • itsnotlupus@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    1 year ago

    I’ll note that there are plenty of models out there that aren’t LLMs and that are also being trained on large datasets gathered from public sources.

    Image generation models, music generation models, etc.
    Heck, it doesn’t even need to be about generation. Music recognition and image recognition models can also be trained on the same sort of datasets, and arguably come with similar IP right questions.

    It’s definitely a broader topic than just LLMs, and attempting to enumerate exhaustively the flavors of AIs/models/whatever that should be part of this discussion is fairly futile given the fast evolving nature of the field.

    • themarty27@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      edit-2
      1 year ago

      Still, all those models are, even conceptually, far removed frow AI. They would most properly be called Machine Learning Models (MLMs).

      • itsnotlupus@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        ·
        1 year ago

        The term AI was coined many decades ago to encompass a broad set of difficult problems, many of which have become less difficult over time.

        There’s a natural temptation to remove solved problems from the set of AI problems, so playing chess is no longer AI, diagnosing diseases through a set of expert system rules is no longer AI, processing natural language is no longer AI, and maybe training and using large models is no longer AI nowadays.

        Maybe we do this because we view intelligence as a fundamentally magical property, and anything that has been fully described has necessarily lost all its magic in the process.
        But that means that “AI” can never be used to label anything that actually exists, only to gesture broadly at the horizon of what might come.

      • gammasfor@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        0
        arrow-down
        1
        ·
        1 year ago

        They would but that doesn’tv sound as sexy to investors.

        That’s what it all comes down to when businesses use words like AI, big data, blockchain etc. Its not about whether it’s an accurate descriptor, its about tricking dumb millionaires into throwing money at them.

  • BURN@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    AI has been a blanket term for Machine Learning, LLMs, Decision Trees and every other form of “intelligence”.

    Unfortunately I think that genie is out of the bottle and it’s never going back in.

    • stewsters@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      And it has been the technical term used in academia since the 1950’s. If anyone is surprised by this usage then they have not studied it, only watched movies.

  • DrQuint@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    If we’re unmasking tech, LLM’s right now are also all just Computer Vision models with a lot of more abstraction layers thrown at them. Nothing but fit assessment machine with a ludicrous amount of extra steps.

    I am convinced this is all pedantry, and these models are going to become the de facto basis for true AI at some point. It was already weird enough that this type of tech got discovered from the goal of checking if an image has a cat or not.

  • Greenskye@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    1 year ago

    We shouldn’t have to pay for the content we use to train and teach an AI

    Wait people think that sounds reasonable?

    • theneverfox@pawb.social
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      I do. If it’s publicly available, individuals should be able to learn from it. Artists don’t pay their influences that helped develop their style, we don’t pay the programmers that answer questions on stack overflow

      Hell, I’m not sure generative AI should have to pay for training data at all. It points to a weakness in the system, and it doesn’t fix it - the field is getting away from needing existing datasets. GPT4 swallowed everything worth swallowing, and it’s already training GPT4.5. This would only make it harder for new players to compete in the generative AI space

      It can’t profit only the few, it’s too big a force multiplier. Paying up front doesn’t fix it, recurring payments don’t fix it… That’s nothing but a payoff to a few people as this starts to eat the best parts of the job market

      We need to think much bigger - we need to look at how we handle ownership as a society

  • lolpostslol@kbin.social
    link
    fedilink
    arrow-up
    0
    arrow-down
    1
    ·
    1 year ago

    It’s just a happy coincidence for them, they call it AI because calling it “a search engine that steals stuff instead of linking to it and blends different sources together to look smarter” wouldn’t be as interesting to clueless financial markets people