I’ve recently noticed this opinion seems unpopular, at least on Lemmy.

There is nothing wrong with downloading public data and doing statistical analysis on it, which is pretty much what these ML models do. They are not redistributing other peoples’ works (well, sometimes they do, unintentionally, and safeguards to prevent this are usually built-in). The training data is generally much, much larger than the model sizes, so it is generally not possible for the models to reconstruct random specific works. They are not creating derivative works, in the legal sense, because they do not copy and modify the original works; they generate “new” content based on probabilities.

My opinion on the subject is pretty much in agreement with this document from the EFF: https://www.eff.org/document/eff-two-pager-ai

I understand the hate for companies using data you would reasonably expect would be private. I understand hate for purposely over-fitting the model on data to reproduce people’s “likeness.” I understand the hate for AI generated shit (because it is shit). I really don’t understand where all this hate for using public data for building a “statistical” model to “learn” general patterns is coming from.

I can also understand the anxiety people may feel, if they believe all the AI hype, that it will eliminate jobs. I don’t think AI is going to be able to directly replace people any time soon. It will probably improve productivity (with stuff like background-removers, better autocomplete, etc), which might eliminate some jobs, but that’s really just a problem with capitalism, and productivity increases are generally considered good.

  • the_toast_is_gone@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    3 months ago

    Wouldn’t that argument be like saying an image I drew of a copyrighted character is just an arrangement of pixels based on existing data? The fact remains that, if I tell an AI to generate an image of a copyrighted character, then it’ll produce something without the permission of the original artist.

    I suppose then the problem becomes, who do you hold responsible for the copyright violation (if you pursue that course of action)? Do you go after the guy who told the AI to do it, or do the people who trained the AI and published it? Possibly both? On one hand, suing the AI AL company would be like suing Adobe because they don’t stop people from drawing copyrighted materials in their software (yet). On the other hand, they did create this software that basically acts in the place of an artist that draws whatever you want for commission. If that artist was drawing copyrighted characters for money, you could make the case that the AI company is doing the same - manufacturing copyrighted character images by feeding the AI images of the character and allowing people to generate images of it while collecting money for their services.

    All this to say, copyright is stupid.