I’m mostly talking about being able to train on copyrighted content. This is on me though, I got mixed up. That’s what I meant in my first comment.
If you think someone can train a model on legally obtained data (Google images, YouTube, internet archive), then that is fair.
Personally, I think using pirated or at least bought content that is ripped (Netflix, DVDs) should be exempt (for everyone obviously, not just OpenAI.) Some data is already behind huge mega corps like record labels, Hollywood, publishing houses, etc. OpenAI can afford the cost but the little guys will be screwed when it comes to SOTA.
It’s also worth noting that most current lawsuits are aimed at how the data is used and not how it’s sourced if I’m not mistaken. The laws coming from these lawsuits won’t be used to bolster anti-piracy laws but copyright laws instead, targeting fair use and transformative clauses imo.
It’s sadly already happening in regards to stack.
Mostly youtube, reddit and image search. I guess I could just record a Netflix stream if I needed the whole movie. I guess recording a Netflix stream is pirating? Probably easier with a torrent.
What does it matters? I don’t think pirating is unethical especially when it’s not even redistribution but transformative. Openai has never stopped me from pirating or even asked me to stop. Not sure what you mean with “no one else”.
You ever ask yourself if the memes made from movie scenes used pirated media?
What pirate bay is doing isn’t exactly transformative. I pirate most of my media and can’t say I’m not for better copyright laws and a better treatment of pirate bay, I just think the situations are different.
I don’t think saying “if pirate bay is illegal, so should training ai without compensations” is exactly fair. (I wish the actual people contributing could be compensated, but how it’s set up, we would be giving a few companies a monopoly while compensating mostly data aggregators.)
Reforms don’t have to be pro-corporate slop.
Sadly, the media and most of the population is practically begging for it. When you couple that with the pressure exerted by record companies, publishing houses, etc, it is clear those are the reforms we get if any.
In our current society, little people can get away with it. I can take whatever style I want and train a model on it. There’s already many ghibli ressources in the open source scene, and a lot of them date from 2 years ago.
This whole situation is rage bait to manipulate the population into cheering for new copyright laws so politicians get little push back when they start writing pro-corporate laws regarding AI.
I understand the sentiment but I think it’s foolhardy.
And all that mostly benefiting the data holders and big ai companies. Most image data is on platforms like Getty, Deviant Art, Instagram, etc. It’s even worse for music and lit, where three record labels and five publishers own most of it.
If we don’t get a proper music model before the lawsuits pass, we will never be able to generate music without being told what is or isn’t okay to write about.
I think it will be punished, but not how we hope. The laws will end up rewarding the big data holders (Getty, record labels, publishers) while locking out open source tools. The paywalls will stay and grow. It’ll just formalize a monopoly.
It shouldn’t be much of a problem using a gibli based model with img2img. I personally use forge as my main ui, models can be found on civitai.com . It’s easily possible, you just need a bit of vram and setting it up is more work. You might get more mileage by using controlnet in conjunction with img2img.
I’m not sure if grok is winning here but Lilly is definitely losing.
The responsable thing would be to put a rule stating her picture and personal information must be blurred. Its a harassment campaign and the community you created is essentially amplifying it.
Same can’t be said for the snacks though. Highway robbery, I tell ya.
The Hollywood business model is coming to an end.
“X”, I only need to tap it three times to get what I want.
Children of time had a lot of this. One factions technology is mostly based on natural processes. Their most complicated computer systems are ant based if I remember well. Great book.
Even jellyfish can learn to spot and dodge obstacles, and they literally don’t have a brain.
AI has a vibrant open source scene and is definitely not owned by a few people.
A lot of the data to train it is only owned by a few people though. It is record companies and publishing houses winning their lawsuits that will lead to dystopia. It’s a shame to see so many actually cheering them on.
I just want the guy managing ressources to not be plowing children.
They won’t be rewarded. Data brokers, record companies, publishing houses, getty, etc will be rewarded.
You want to shoot open source initiatives in the face and give a handful of companies a monopoly, so rich people can get richer.
seeing the huge amount of data needed for competitive generative AI, then open source AI cannot afford the data and dies.
It isn’t. I’d even say that simply completing puzzles is far from AGI, even if the puzzles are complex.