this post was submitted on 30 Jul 2023
224 points (100.0% liked)

Technology

37747 readers
212 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

Greg Rutkowski, a digital artist known for his surreal style, opposes AI art but his name and style have been frequently used by AI art generators without his consent. In response, Stable Diffusion removed his work from their dataset in version 2.0. However, the community has now created a tool to emulate Rutkowski's style against his wishes using a LoRA model. While some argue this is unethical, others justify it since Rutkowski's art has already been widely used in Stable Diffusion 1.5. The debate highlights the blurry line between innovation and infringement in the emerging field of AI art.

you are viewing a single comment's thread
view the rest of the comments
[–] FaceDeer@kbin.social 11 points 1 year ago (1 children)

They can't, and "intact" watermarks don't show up. You're the one who is misunderstanding how this works.

When a pattern is present very frequently the AI can learn to imitate it, resulting in things that closely resemble known watermarks. This is called "overfitting" and is avoided as much as possible. But even in those cases, if you examine the watermark-like pattern closely you'll see that it's usually quite badly distorted and only vaguely watermark-like.

[–] Pulse@dormi.zone 11 points 1 year ago (1 children)

Yes, because "imitate" and "copy" are different things when stealing from someone.

I do understand how it works, the "overfitting" was just laying clear what it does. It copies but tries to sample things in a way that won't look like clear copies. It had no creativity, it is trying to find new ways of making copies.

If any of this was ethical, the companies doing it would have just asked for permission. That they didn't says a everything you need to know.

I don't usually have these kinds discussions anymore, I got tired of conversations like this back in 2016, when it became clear that people will go to the ends of the earth to justify unethical behavior as long as the people being hurt by it are people they don't care about.

[–] FaceDeer@kbin.social 5 points 1 year ago (1 children)

And we're back to you calling it "stealing", which it certainly is not. Even if it was copyright violation, copyright violation is not stealing.

You should try to get the basic terminology right, at the very least.

[–] Pulse@dormi.zone 6 points 1 year ago (1 children)

Just because you've redefined theft in a way that makes you feel okay about it doesn't change what they did.

They took someone else's work product, fed it into their machine then used that to make money.

They stole someone's labor.

[–] FaceDeer@kbin.social 4 points 1 year ago (1 children)

I haven't "redefined" it, I'm using the legal definition. People do sometimes sloppily equate copyright violation with theft in common parlance, but they're in for a rude awakening if they intend to try translating that into legal action.

Using that term in an argument like this is merely trying to beg the question of whether it's wrong, since most everyone agrees that stealing is wrong you're trying to cast the action of training an AI as something everyone will by default agree is wrong. But it's not stealing, no matter how much you want it to be, and I'm calling that rhetorical trick out here.

If you want to argue that it's wrong you need to argue against the actual process that's happening, not some magical scenario where the AI trainers are somehow literally robbing people.

[–] Pulse@dormi.zone 3 points 1 year ago (1 children)

Taking someone's work product and converting it, without compensation and consent, into your profit is theft of labor.

Adding extra steps, like, say, training an AI, doesn't absolve the theft of labor.

We're it ethical, the companies doing it would have asked for permission and been given cinsent. They didn't.

[–] FaceDeer@kbin.social 3 points 1 year ago (1 children)

Taking someone’s work product and converting it, without compensation and consent, into your profit is theft of labor.

That's not what's going on here. The finished product contains only the style of the artist that the AI was trained on, and style is not copyrightable. Which is a damn good thing, as humans have been learning from each other's "work products" and mimicking each others' styles since time immemorial.

BTW, theft of labor means failing to pay wages or provide employee benefits owed to an employee by contract or law. You're using that term incorrectly too, Greg Rutkowski wasn't hired to do anything for the people who trained the AI off of his work.

[–] Pulse@dormi.zone 4 points 1 year ago (1 children)

No, I'm not using it incorrectly, I'm just not concerned with the legal definition as I'm not a lawyer or anyone tied up in this mess.

If you do a thing, and it takes time and skill to do it, then someone copies it, they stole your labor.

Saying they "copied his style", the style he spent a lifetime crafting, then trying to say they didn't benefit, at no cost, to the labor he put into crafting that style because "well actually, the law says..." is a bad argument as it tries to minimize what they did.

If their product could not exist without his labor, and they did not pay him for that labor, they stole his labor.

For, like, the fourth time in this thread: were this ethical, they would have asked for permission, they didn't.

[–] FaceDeer@kbin.social 2 points 1 year ago (1 children)

If you're just going to make up the meanings of words there's not much point in using them any further.

[–] Pulse@dormi.zone 3 points 1 year ago (1 children)

But I'm not.

You're trying to say that, because this one law doesn't say it's bad it must therefore be good (or at least okay).

I'm simply saying that if you profit from someone else's labor, without compensating them (or at least getting their consent), you've stolen the output of that labor.

I'm happy to be done with this, I didn't expect my first Lemmy comment to get any attention, but no, I'm not going to suddenly be okay with this just because the legal definition of "stealing labor" is to narrow to fit this scenario.

[–] whelmer@beehaw.org 1 points 1 year ago

The law doesn't even say it's okay. What FaceDeer is referring to is that copyright infringement is a different category of crime than theft, which is defined as pertaining to physical property. It's a meaningless point because, as you said, this isn't a courtroom and we aren't lawyers and the concept of intellectual property theft is well understood.

It's a thing engineers and lawyers often seem to do, to take the way terms are used in a particular professional jargon and assume that that usage is "the real" usage.