this post was submitted on 22 Dec 2024
1296 points (97.3% liked)

Technology

60055 readers
3090 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
 

It's all made from our data, anyway, so it should be ours to use as we want

you are viewing a single comment's thread
view the rest of the comments
[–] mp3@lemmy.ca 38 points 20 hours ago (4 children)

It could also contain non-public domain data, and you can't declare someone else's intellectual property as public domain just like that, otherwise a malicious actor could just train a model with a bunch of misappropriated data, get caught (intentionally or not) and then force all that data into public domain.

Laws are never simple.

[–] drkt@scribe.disroot.org 17 points 20 hours ago (1 children)

Forcing a bunch of neural weights into the public domain doesn't make the data they were trained on also public domain, in fact it doesn't even reveal what they were trained on.

[–] grue@lemmy.world 13 points 20 hours ago (1 children)

So what you're saying is that there's no way to make it legal and it simply needs to be deleted entirely.

I agree.

[–] FaceDeer@fedia.io 4 points 18 hours ago (2 children)

There's no need to "make it legal", things are legal by default until a law is passed to make them illegal. Or a court precedent is set that establishes that an existing law applies to the new thing under discussion.

Training an AI doesn't involve copying the training data, the AI model doesn't literally "contain" the stuff it's trained on. So it's not likely that existing copyright law makes it illegal to do without permission.

[–] grue@lemmy.world 1 points 6 hours ago

There’s no need to “make it legal”, things are legal by default until a law is passed to make them illegal.

Yes, and that's already happened: it's called "copyright law." You can't mix things with incompatible licenses into a derivative work and pretend it's okay.

[–] xigoi@lemmy.sdf.org 1 points 15 hours ago (3 children)

By this logic, you can copy a copyrighted imege as long as you decrease the resolution, because the new image does not contain all the information in the original one.

[–] Voyajer@lemmy.world 2 points 13 hours ago

More like reduce it to a handful of vectors that get merged with other vectors.

[–] yetAnotherUser@discuss.tchncs.de 1 points 11 hours ago (1 children)

Am I allowed to take a copyrighted image, decrease its size to 1x1 pixels and publish it? What about 2x2?

It's very much not clear when a modification violates copyright because copyright is extremely vague to begin with.

[–] grue@lemmy.world 3 points 6 hours ago (1 children)

Just because something is defined legally instead of technologically, that doesn't make it vague. The modification violates copyright when the result is a derivative work; no more, no less.

[–] yetAnotherUser@discuss.tchncs.de 1 points 2 hours ago (1 children)

What is a derivative work though? That's again extremely vague and has been subject to countless lawsuits seeking to determine the bounds.

[–] catloaf@lemm.ee 2 points 1 hour ago

If your work depends on the original, such that it could not exist without it, it's derivative.

I can easily create a pixel of any arbitrary color, so it's sufficiently transformative that it's considered a separate work.

The four fair use tests are pretty reliable in making a determination.

[–] FaceDeer@fedia.io -1 points 12 hours ago (1 children)

In the case of Stable Diffusion, they used 5 billion images to train a model 1.83 gigabytes in size. So if you reduce a copyrighted image to 3 bits (not bytes - bits), then yeah, I think you're probably pretty safe.

[–] xigoi@lemmy.sdf.org 1 points 33 minutes ago

Your calculation is assuming that the input images are statistically independent, which is certainly not the case (otherwise the model would be useless for generating new images)

[–] merc@sh.itjust.works 7 points 18 hours ago

It wouldn't contain any public-domain data though. That's the thing with LLMs, once they're trained on data the data is gone and just added to the series of weights in the model somewhere. If it ingested something private like your tax data, it couldn't re-create your tax data on command, that data is now gone, but if it's seen enough private tax data it could give something that looked a lot like a tax return to someone with an untrained eye. But, a tax accountant would easily see flaws in it.

[–] pelespirit@sh.itjust.works 3 points 20 hours ago