Anyone can copy it, recreate with it, reproduce with it
Ew… stay away from my content, you creep!
Anyone can copy it, recreate with it, reproduce with it
Ew… stay away from my content, you creep!
If you think of LLMs as being akin to lossy text compression of a set of text, where the compression artifacts happen to also result in grammatical-looking sentences, the question you eventually end up asking is “why is the compression lossy? What if we had the same thing but it returned text from its database without chewing it up first?” and then you realize that you’ve come full circle and reinvented search engines
unironically saying “the sharing economy” in the year of our lord 2024 is… certainly a choice
also
God knows we old-timers tried to be cynical about ChatGPT, pedantically insisting that AI was actually just machine learning and that Altman’s new toy was nothing but cheap mimicry. But the rest of the world knew better
idk dude I’ve talked to the rest of the world about this and most of them actually seem to dislike this technology, it seems like maybe you didn’t actually try very hard to be cynical
The copyright clause in the US constitution (1789) also frames it in terms of granting rights to authors to “promote the progress of … useful arts”. Strictly speaking author protection is not the origin of copyright but also I was snarkily responding to a person who was arguing in favor of AI-training-as-fair-use and implying copyright was 120 years old, not trying to do a detailed explication of the origins of copyright law
I’m sorry for my imprecise wording, I was feeling flippant and I know what I said isn’t totally accurate. not a big history person here honestly. I’ll try and stick to joke-commenting next time. but also can you just say what you mean instead of darkly hinting.
iirc even though the origin of copyright is not really specifically about author protection, part of the broad-strokes motivation for its existence involved “we need to keep production of new works viable in a world where new copies can be easily produced and undercut the original,” which was what I was trying to get at. maybe they picked a bad way to do that idk I’m not here to make excuses for the decisions of 16th-century monarchs
also again I’m not a copyright fan/defender. in particular copyright as currently constituted massively and obviously sucks. I just don’t think copyright-in-the-abstract is like the Greatest Moral Evil either, bc I’m not a libertarian. sorry ¯\_(ツ)_/¯
heck yeah I love Physics Jenny Nicholson Angela Collier
I mean, it seems like you’re reading my argument as a defense of copyright as a concept. I’m ambivalent on the goodness or badness of copyright law in the abstract. Like a lot of laws, it’s probably not the ideal way to fix the issue it was designed to solve, and it comes with (many) issues of its own, but that doesn’t necessarily mean we’d be better off if we just got rid of it wholesale and left the rest of society as is. (We would probably be left with excitingly new and different problems.)
As I see it, the actual issue at hand with all of this is that people are exploiting the labor/art/culture of others in order to make a profit for themselves at the expense of the people affected. Sometimes copyright is a tool to facilitate that exploitation, and sometimes it’s a tool that protects people from it. To paraphrase Dan Olson, the problem is what people are doing to others, not that the law they’re using to do it is called “copyright.”
That seems bad but also not super relevant to the point under discussion! Unless your point is that it’s bad when a cultural commons is exploited for business profits – in which case, I agree, but, well…
Haha, sounds like we might have to agree to disagree on this one.
Copyright is much older than 1904, though! It dates back to the printing press, when it became necessary because the new technology made it possible to benefit off writers’ work without compensating them, which made it hard to be a writer as a profession, even though we want people to be able to do that as a society. Hey, wait a minute…
Yeah but this presumes “the best way to beat 'em is to join 'em,” right? Like, when all the operating systems or databases are proprietary, that’s bad because those things are really useful and help you do things better and faster than you would otherwise.
But this argument applied here is like, oh no, what if large entertainment companies start making all their movies out of AI garbage, and everyone else can’t do that because they can’t get the content licensed? Well… what if they do? Does that mean they’re going to be making stuff that’s better? Wouldn’t the best way to compete with that be not to use the technology because you’ll get a higher-quality product? Or are we just giving up on the idea of producing good art at all and conceding that yes we actually only value cheapness and quantity?
Also, just on a personal level, for me as a J. Random Person who uploads creative work to the internet (some of which is in common crawl), but who doesn’t work for a major entertainment corporation that has rights to my work, I would really prefer to have a way to say “sorry no, you can’t use my stuff for this.” I don’t really find “well you see, we need to be able to compete with large entertainment companies in spam content generation, so we need to be able to use your uncompensated labor for our benefit without your permission and without crediting you” particularly compelling.
Yeah, I think his ideological commitment to “all intellectual property rights are bad forever and always amen” kind of blinds him to the actual issue here, and his proposed solution is kind of nonsensical in terms of its ability to get off the ground.
More broadly, (ie not just in relation to Cory Doctorow), I’ve seen the take floating around that’s like “hey, what the heck, artists who were opposed to ridiculous IP rights restrictions when it was the music industry doing it are now in favor of those restrictions when it’s AI, what gives with this hypocrisy?” which I think kind of… misses the point?
A lot of artists generally are in favor of using their work for interesting collaborative stuff and aren’t going to get mad if you use their stuff for your own creative endeavors. This is why we have things like Creative Commons. The actual things artists tend not to like are things like having their work used for commercial purposes without permission and/or having their work taken without credit. (This is why CC licenses often restrict these usages!) With that in mind, a lot of the artist outrage over AI feels much more in line with artists getting mad about, say, watermark-removal tools, or people reposting art without credit, than it does with the copyright battles of the 00s. (You may remember one of the big things artists were affronted by about AI art was the way it would imitate an artist’s signature, because of what that represented.)
In this case, artists are leaning on copyright not out of any particular ideological commitment but just because it’s the blunt instrument that they already have at their disposal. But I think Cory Doctorow’s previous experience in “getting mad at the MPAA” or whatever kind of forces him to analyze this using the same framing as that issue, which doesn’t really make sense in this case. And ironically saying “copyright shouldn’t count for AI” aligns him with the position of the MPAA so it really does feel like a “live long enough to see yourself become the villain” scenario. :/
hold on, when did the “first generation” of generative ai start?
I believe waitbutwhy came up before on old sneerclub though in that case we were making fun of them for bad political philosophy rather than bad ai takes
Making me learn how to do things the right way is premature optimization
Ah yes, pragmatists, well known for their constantly sunny and optimistic outlook on the future, consequences be damned (?)
No no, it’s “order of magnitudes”. It’s like “surgeons general.”
This is my favorite LLM response from the paper I think:
It’s really got everything – they surrounded the problem with the recommended prompt engineering garbage, which results in the LLM first immediately directly misstating the prompt, then making a logic error on top of that incorrect assumption. Then when it tries to consider alternate possibilities it devolves into some kind of corporate-speak nonsense about ‘inclusive language’, misinterprets the phrase ‘inclusive language’, gets distracted and starts talking about gender identity, then makes another reasoning error on top of that! (Three to five? What? Why?)
And then as the icing on the cake, it goes back to its initial faulty restatement of the problem and confidently plonks that down as the correct answer surrounded by a bunch of irrelevant waffle that doesn’t even relate to the question but sounds superficially thoughtful. (It doesn’t matter how many of her nb siblings might identify as sisters because we already know exactly how many sisters she has! Their precise gender identity completely doesn’t matter!)
Truly a perfect storm of AI nonsense.
Q: When you think about the big vision — which still my mind is blown that this is your big vision, — of “I’m going to send a digital twin into a meeting, and it’s going to make decisions on my behalf that everyone trusts, that everyone agrees on, and everyone acts upon,” the privacy risk there is even higher. The security surface there becomes even more ripe for attack. If you can hack into my Zoom and get my digital twin to go do stuff on my behalf, woah, that’s a big problem. How do you think about managing that over time as you build toward that vision?
A: That’s a good question. So, I think again, back to privacy and security, I think of two things. First of all, it’s how to make sure somebody else will not hack into your meeting. This is Eric; it’s not somebody else. Another thing: during the call, make sure your conversation is very secure. Literally just last week, we announced the industry’s first post-quantum encryption. That’s the first one, and at the same time, look at deepfake technology — we’re also working on that as well to make sure that deepfakes will not create problems down the road. It is not like today’s two-factor authentication. It’s more than that, right? And because deepfake technology is real, now with AI, this is something we’re also working on — how to improve that experience as well.
Spoken like a true person who has not given one iota of thought to this issue and doesn’t know what most of the words he’s saying mean
Wow, this comment definitely caught my attention! “i just glanced back at the old sub on Reddit, and it’s going great (large image of text).” Sounds like the old sub on Reddit is going great! It reminds me of how people post on Reddit about things. I’m curious to hear what’s in the large image of text. Have any of you ever checked old subs on Reddit? How were they going? Let’s dive into this intriguing topic together!
ngl his stuff always felt a bit cynical to me, in that it seemed to exist more to say “look, video games can have a deep message!” than it did to just have such a message in the first place. Like it existed more to gesture at the concept of meaningfulness rather than to be meaningful itself.