Don’t forget we moved!
https://brandmu.day/
AI Megathread
-
As not to pollute the other thread, in regards to MUs becoming heavily reliant on AI:
The official RPGs are doing it, so why shouldn’t you?! Hasbro has announced AI DMs as an upcoming feature for D&D Beyond, and they were just caught using AI generated art in their newest upcoming/just releasing book, with some pretty blatantly poor looking results.
That’s to say nothing of the major film industry strikes right now being triggered by studios wanting to replace writers with AI and digitally catalogue extras so they can deepfake them for eternity.
Thrash and struggle as we may, but this is the world we live in now.
-
@bored said in AI Megathread:
Thrash and struggle as we may, but this is the world we live in now.
Only if we accept it.
The actors and writers in Hollywood are striking because of it.
Lots of folks in the writing community are boycotting AI-generated covers or ChatGPT-generated text, and there’s backlash against those who use them.
Lawsuits are striking back against the copyright infringement.
Again, to be clear, I’m not against the underlying technology, only the unethical use of it.
If a specific author or artist wants to train a model on their own stuff and then use it to generate more stuff like their own? More power to them. (It won’t work as well, because the real horsepower comes from the sheer volume of trained work, but that’s a separate issue.)
If MJ were only trained on a database of work from artists who had opted in and were paid royalties for every image generated? (like @Testament mentioned for Shutterstock - 123rf and Adobe have similar systems) That’s fine too (assuming the royalty arrangements are decent - look to Spotify for the dangers there).
These tools are products, and consumers have an influence in whether those products are commercially successful.
-
@Faraday There will be a legal shake-out, for sure. I’m not optimistic that starving artists (a group not traditionally known for their ability to afford expensive lobbyists) are going to win, though. While Midjourney is a bit of a black box (given their for-profit model), the idea of figuring out valid royalties for Stable Diffusion’s training data is getting into counting grains of sand on the beach territory. Given the open-source nature and proliferation of descendant models… what can you do? The only answer is to ban the technology completely outside of… approved, licensed models, which would almost certainly just be MORE of a corporate coup as ownership of those big data sets will become prized.
I brought this one up because I think the Hasbro stuff is an interesting case of it being mainstreamed and placed into the professional space (And its RPGs. Exactly what we discuss here.) Notably, the book has artist credits! These are a bunch of regular WotC contributors, and the art is in familiar styles (although also, the artist whose work it most looks like is not in the credits). Something got arranged here. Some people got paid. But maybe not all the people. WoTC presumably owns the rights to every bit of D&D and (vastly more) MTG art. That is a big enough set to train on. They’re probably not paying most of those people.
There will be laws for all this stuff eventually, but the idea that the creative industries come out on top seems very slim. Copywriting is already essentially being annihilated as a profession. I don’t see how it goes any other way.
-
@bored said in AI Megathread:
I’m not optimistic that starving artists (a group not traditionally known for their ability to afford expensive lobbyists) are going to win, though.
Not directly, no, but it’s not only starving artists fighting. Big-name companies like Getty Images are leading the charge, and they could get protections in place that impact the little guys too.
@bored said in AI Megathread:
the idea of figuring out valid royalties for Stable Diffusion’s training data is getting into counting grains of sand on the beach territory.
Sure, but that’s kind of the point. If a lawsuit decided that this was all copyright infringement and they were required to pay royalties to artists whose work was used in training, they’d basically have no choice but to throw away the existing models and start with more ethical ones. Would the old ones still exist in rogue form on the internet? Sure, but they would cease being in the mainstream.
Book piracy and music/movie torrent sites didn’t vanish after the fall of Napster, but nobody tried to make it their business model.
-
I agree large rightsholders pushing is the only way those legal chances happen. I am extremely dubious that this somehow yields an ‘ethical’ outcome. Are we fans of the DMCA now?
If generative AI becomes subject to proof of ownership of your training data, then it will just become de facto the exclusive tool of large tech and media companies that already own that data (or who can encourage you to sign away your rights to it as a part of their services, or just purchase it). Subsequently, everyone who wants to use it will be paying for their services. Google and Meta and Disney win, not John StockModel or Jane FanArtist. At best, they’ll be given the option to sign over their work forever, for very small payouts.
And it’s fundamentally worse than any of the analogy cases because once you have your model, you genuinely never need those models or artists again. That’s why the tech is worth it. It’s pointless if you owe royalties. Even if it wins a court case, Getty will lose (or realistically, sell the data while it’s still worth anything): Google scanned the planet, it can generate new training data if it has to. The actors strike will settle, they’ll give up their Bond villain position as a negotiating point and just generate novel fake humans instead of buying someone’s face. The tech is so transformative that anyone who thinks they can fight it is just going to get run over.
-
@bored said in AI Megathread:
The actors strike will settle, they’ll give up their Bond villain position as a negotiating point and just generate novel fake humans instead of buying someone’s face
What will they use to generate those fake humans
Eta:
@bored said in AI Megathread:
Google scanned the planet, it can generate new training data if it has to
This is way more legally fraught than you’d think, especially given European privacy rights
-
@Rinel said in AI Megathread:
@bored said in AI Megathread:
The actors strike will settle, they’ll give up their Bond villain position as a negotiating point and just generate novel fake humans instead of buying someone’s face
What will they use to generate those fake humans
There’s already plenty of those on TikTok.
-
@Testament said in AI Megathread:
@Rinel said in AI Megathread:
@bored said in AI Megathread:
The actors strike will settle, they’ll give up their Bond villain position as a negotiating point and just generate novel fake humans instead of buying someone’s face
What will they use to generate those fake humans
There’s already plenty of those on TikTok.
In that case the actors are safe
-
@Faraday said in AI Megathread:
@bored said in AI Megathread:
Thrash and struggle as we may, but this is the world we live in now.
Only if we accept it.
The actors and writers in Hollywood are striking because of it.
Lots of folks in the writing community are boycotting AI-generated covers or ChatGPT-generated text, and there’s backlash against those who use them.
Lawsuits are striking back against the copyright infringement.
Again, to be clear, I’m not against the underlying technology, only the unethical use of it.
If a specific author or artist wants to train a model on their own stuff and then use it to generate more stuff like their own? More power to them. (It won’t work as well, because the real horsepower comes from the sheer volume of trained work, but that’s a separate issue.)
If MJ were only trained on a database of work from artists who had opted in and were paid royalties for every image generated? (like @Testament mentioned for Shutterstock - 123rf and Adobe have similar systems) That’s fine too (assuming the royalty arrangements are decent - look to Spotify for the dangers there).
These tools are products, and consumers have an influence in whether those products are commercially successful.
Right now at work I’m having a horrendous time dealing with upper management who are just uncritical (and frankly, experienced enough to know better) about this sort of thing.
“ChatGPT told <Team Member x> to do thing Y with Python! It even wrote them the script!”
Me: Uhh, do they have any idea how this works? Or how to even read the script?
Dead silence.
I’m not letting this go on in my org without a fight.
-
@bored said in AI Megathread:
@Faraday There will be a legal shake-out, for sure. I’m not optimistic that starving artists (a group not traditionally known for their ability to afford expensive lobbyists) are going to win, though. While Midjourney is a bit of a black box (given their for-profit model), the idea of figuring out valid royalties for Stable Diffusion’s training data is getting into counting grains of sand on the beach territory. Given the open-source nature and proliferation of descendant models… what can you do? The only answer is to ban the technology completely outside of… approved, licensed models, which would almost certainly just be MORE of a corporate coup as ownership of those big data sets will become prized.
I brought this one up because I think the Hasbro stuff is an interesting case of it being mainstreamed and placed into the professional space (And its RPGs. Exactly what we discuss here.) Notably, the book has artist credits! These are a bunch of regular WotC contributors, and the art is in familiar styles (although also, the artist whose work it most looks like is not in the credits). Something got arranged here. Some people got paid. But maybe not all the people. WoTC presumably owns the rights to every bit of D&D and (vastly more) MTG art. That is a big enough set to train on. They’re probably not paying most of those people.
There will be laws for all this stuff eventually, but the idea that the creative industries come out on top seems very slim. Copywriting is already essentially being annihilated as a profession. I don’t see how it goes any other way.
Y’all ever think all of this shit is just incompatible with capitalism?
Or rather, ever think that capitalism is just incompatible with life at this point?
I’ll cop to oversimplifying things and thumping Marx and Marcuse here, but IMO the quiet part hasn’t been said loud enough in all of this mess.
-
@bored said in AI Megathread:
Are we fans of the DMCA now?
There’s certainly room to improve copyright laws, but am I a fan of the broad principles behind it? Of protecting creators’ work? Absolutely.
Creators are going nowhere because large language models and stable diffusion type graphic models are dumb.
I don’t just mean they’re dumb on principle, I mean they’re dumb logically and artistically. They’re parrots who don’t understand the world they’re parroting. They don’t understand jack about the human experience.
They can never write a news article about an event that just happened - not until some human writes it first so they can copy it. They can’t write a biography about a person who hasn’t been born yet, or generate an image of Ford’s new automobile, or write late-night jokes about the day’s events - not until a human has done something for them to copy. They will not - and can not - ever generate something truly original, inspired, or new.
Now to be fair, a lot of human creations are unoriginal and uninspired too, but we at least have the potential to do better. Generative AI doesn’t.
And this isn’t just a bug that they can fix in version 2.0. It’s baked into the very core of how they work.
JMS had a great essay on this.
I also liked Adam Conover’s video essay.
I don’t mean to downplay the harm that generative AI tools can cause to creatives. But this tech is nowhere near as transformative as the tech hype would like us to believe.
-
@SpaceKhomeini said in AI Megathread:
Y’all ever think all of this shit is just incompatible with capitalism?
The unrestrained drive for ever increasing profit will result in reform, revolution, or societal collapse. So yes.
You could probably keep capitalism going indefinitely even with shit like this if you implemented a strong safety program, but the plutocrats are too stupid even to do that. So!
-
I think we’re just going to see writing and graphical art go the same way as physical art that can be mass-produced. I can get an imitation porcelain vase mass produced out of a factory for $10 or I can get a handmade artisan-crafted porcelain vase for $1000. They may seem alike but a critical eye can spot the difference and some people may care about that.
-
@SpaceKhomeini said in AI Megathread:
@Faraday said in AI Megathread:
@bored said in AI Megathread:
Thrash and struggle as we may, but this is the world we live in now.
Only if we accept it.
The actors and writers in Hollywood are striking because of it.
Lots of folks in the writing community are boycotting AI-generated covers or ChatGPT-generated text, and there’s backlash against those who use them.
Lawsuits are striking back against the copyright infringement.
Again, to be clear, I’m not against the underlying technology, only the unethical use of it.
If a specific author or artist wants to train a model on their own stuff and then use it to generate more stuff like their own? More power to them. (It won’t work as well, because the real horsepower comes from the sheer volume of trained work, but that’s a separate issue.)
If MJ were only trained on a database of work from artists who had opted in and were paid royalties for every image generated? (like @Testament mentioned for Shutterstock - 123rf and Adobe have similar systems) That’s fine too (assuming the royalty arrangements are decent - look to Spotify for the dangers there).
These tools are products, and consumers have an influence in whether those products are commercially successful.
Right now at work I’m having a horrendous time dealing with upper management who are just uncritical (and frankly, experienced enough to know better) about this sort of thing.
“ChatGPT told <Team Member x> to do thing Y with Python! It even wrote them the script!”
Me: Uhh, do they have any idea how this works? Or how to even read the script?
Dead silence.
I’m not letting this go on in my org without a fight.
Please tell me you’re documenting this, because this is a security breach waiting to happen, or other litigation.
-
@dvoraen said in AI Megathread:
@SpaceKhomeini said in AI Megathread:
@Faraday said in AI Megathread:
@bored said in AI Megathread:
Thrash and struggle as we may, but this is the world we live in now.
Only if we accept it.
The actors and writers in Hollywood are striking because of it.
Lots of folks in the writing community are boycotting AI-generated covers or ChatGPT-generated text, and there’s backlash against those who use them.
Lawsuits are striking back against the copyright infringement.
Again, to be clear, I’m not against the underlying technology, only the unethical use of it.
If a specific author or artist wants to train a model on their own stuff and then use it to generate more stuff like their own? More power to them. (It won’t work as well, because the real horsepower comes from the sheer volume of trained work, but that’s a separate issue.)
If MJ were only trained on a database of work from artists who had opted in and were paid royalties for every image generated? (like @Testament mentioned for Shutterstock - 123rf and Adobe have similar systems) That’s fine too (assuming the royalty arrangements are decent - look to Spotify for the dangers there).
These tools are products, and consumers have an influence in whether those products are commercially successful.
Right now at work I’m having a horrendous time dealing with upper management who are just uncritical (and frankly, experienced enough to know better) about this sort of thing.
“ChatGPT told <Team Member x> to do thing Y with Python! It even wrote them the script!”
Me: Uhh, do they have any idea how this works? Or how to even read the script?
Dead silence.
I’m not letting this go on in my org without a fight.
Please tell me you’re documenting this, because this is a security breach waiting to happen, or other litigation.
At work we were recently talking about AI Hallucination Exploits. Essentially, sometimes AI-generated code calls imports for nonexistent libraries. Attackers can upload so-named libraries to public repositories, full of exploits. Hilarity and bottom-line-affecting events ensue.
-
@shit-piss-love said in AI Megathread:
I think we’re just going to see writing and graphical art go the same way as physical art that can be mass-produced. I can get an imitation porcelain vase mass produced out of a factory for $10 or I can get a handmade artisan-crafted porcelain vase for $1000. They may seem alike but a critical eye can spot the difference and some people may care about that.
As I just mentioned off of this thread, I’m from a family of musicians (only ever a hobbyist myself, but still). The idea that you can make close to a living as one is rather laughable at this point. The apocalypse in this field already happened long before The Zuck opened up his magic generative AI music box.
A myriad of factors already killed your average working musician’s ability to make a living (you’re probably not going to be Bjork, and the days of well-compensated session musicians are kind of gone).
And just like Hollywood, the music industry is already figuring out how to use AI so they never have to sign or employ anyone ever again. But even without AI, the vast majority of commercially-viable pop music being written by the same couple of middle-aged Swedish dudes sort of had the same effect.
For sure, generative AI as a replacement for real artistry is a grotesque insult, but until we start deciding that artists and people in general should be allocated resources to have their needs met regardless of the commercial viability and distribution of their product, we’re just going to be kicking this can down the road.
-
@dvoraen said in AI Megathread:
@SpaceKhomeini said in AI Megathread:
@Faraday said in AI Megathread:
@bored said in AI Megathread:
Thrash and struggle as we may, but this is the world we live in now.
Only if we accept it.
The actors and writers in Hollywood are striking because of it.
Lots of folks in the writing community are boycotting AI-generated covers or ChatGPT-generated text, and there’s backlash against those who use them.
Lawsuits are striking back against the copyright infringement.
Again, to be clear, I’m not against the underlying technology, only the unethical use of it.
If a specific author or artist wants to train a model on their own stuff and then use it to generate more stuff like their own? More power to them. (It won’t work as well, because the real horsepower comes from the sheer volume of trained work, but that’s a separate issue.)
If MJ were only trained on a database of work from artists who had opted in and were paid royalties for every image generated? (like @Testament mentioned for Shutterstock - 123rf and Adobe have similar systems) That’s fine too (assuming the royalty arrangements are decent - look to Spotify for the dangers there).
These tools are products, and consumers have an influence in whether those products are commercially successful.
Right now at work I’m having a horrendous time dealing with upper management who are just uncritical (and frankly, experienced enough to know better) about this sort of thing.
“ChatGPT told <Team Member x> to do thing Y with Python! It even wrote them the script!”
Me: Uhh, do they have any idea how this works? Or how to even read the script?
Dead silence.
I’m not letting this go on in my org without a fight.
Please tell me you’re documenting this, because this is a security breach waiting to happen, or other litigation.
Here’s the fun part where I tell you that one of the clueless people involved has a cybersec job title.
-
-
@SpaceKhomeini said in AI Megathread:
For sure, generative AI as a replacement for real artistry is a grotesque insult, but until we start deciding that artists and people in general should be allocated resources to have their needs met regardless of the commercial viability and distribution of their product, we’re just going to be kicking this can down the road.
Absolutely. I’m in no way endorsing the destructive effects AI is going to have. I just don’t think you can put the genie back in the bottle with technology. There doesn’t seem to be a good way out with our economic model. We could be living in a world filled with so much more beauty and culture.
-
@Rinel They don’t have to steal data. They can just… get people to consent to scans (which will be far more exhaustive and better for training than random scraping). Maybe even pay them a pittance!
The point was that they have the resources to do it, and are simultaneously one of the major movers behind the technology. The idea that the data necessary to generate convincing people is hard to acquire is just… it’s not The movie industry ‘scan extras’ thing was being done for other reasons, not because it’s hard to find humans. I would assume they probably already have it available; Google’s AI models are not based on the broad, wild-west scraping that Stable Diffusion is.
@SpaceKhomeini Yes. If people are reading this as… pro big business, I have no idea how.
@Faraday said in AI Megathread:
@bored said in AI Megathread:
Are we fans of the DMCA now?
There’s certainly room to improve copyright laws, but am I a fan of the broad principles behind it? Of protecting creators’ work? Absolutely.
Principles are great. But what actually happens when the government intervenes to protect stakeholders from technology? At no point is the answer ‘the little guy profits.’ (See, for instance, DMCAs against youtubers).
I don’t just mean they’re dumb on principle, I mean they’re dumb logically and artistically. They’re parrots who don’t understand the world they’re parroting. They don’t understand jack about the human experience.
Be that as it may, artists are clearly threatened by them. And I can see why.
Working with a home install on consumer-grade gaming hardware, I can with a bit of learning and small amount of effort and generate images that are arguably more aesthetically pleasing than most amateur humans, in what is probably a hundredth of the time, with much closer control than I’d get trying to communicate needs to an artist and going through a revision process. For me, this is just fun to play around with, yet I see how that could be immensely valuable to a lot of people. AI can’t replace top 20% human creative talent, maybe, but I think it will obliterate vast midrange of it. I started this tangent with the largest company in our hobby selling a 60 dollar hardcover book with AI art. That’s where we’re at. I thought it was interesting, anyway.