
Microsoft is wasting their money on this.
Microsoft is wasting their money on this.
An alluring oversimplification, but your cynical framework can’t account for the communities of people who put in time and effort into FOSS projects. The quality and popularity of open source alternatives has eroded the moats of proprietary services, making it impossible for them to monopolize and profit from this public technology. So if it happened the way it did, it wasn’t for the reasons stated.
I don’t think there is one single purpose. That’s a hard question to answer.
AI training isn’t only for mega-corporations. Setting up barriers that only benefit the ultra-wealthy will only end with corporations gaining a monopoly of a public technology by making it prohibitively expensive for regular folks. And that’s before they bind users to predatory ToS, allowing them exclusive access to user data and effectively selling our own data back to us. What some people want would mean the end of open access to competitive, corporate-independent tools and would imperil research, reviews, reverse engineering, and even indexing information.
They want you to believe that analyzing things without permission somehow goes against copyright, when in reality, fair use is a part of copyright law, and the reason our discourse isn’t wholly controlled by mega-corporations and the rich. The same people who abuse DMCA takedown requests for their chilling effects on fair use content now need your help to do the same thing to open source AI. Their next greatest foe after libraries, students, researchers, and the public domain. Don’t help them do it.
I recommend reading this article by Cory Doctorow, and this open letter by Katherine Klosek, the director of information policy and federal relations at the Association of Research Libraries.
I’m not discussing the use of private data, nor was I ever. You’re presenting a false Dichotomy and trying to drag me into a completely unrelated discussion.
As for your other point. The difference between this and licensing for music samples is that the threshold for abuse is much, much lower. We’re not talking about hindering just expressive entertainment works. Research, reviews, reverse engineering, and even indexing information would be up in the air. This article by Tori Noble a Staff Attorney at the Electronic Frontier Foundation should explain it better than I can.
Private conversations are something entirely different from publically available data, and not really what we’re discussing here. Compensation for essentially making observations will inevitably lead to abuse of the system and deliver AI into the hands of the stupidly rich, something the world doesn’t need.
I mean realistically, we don’t have any proper rules in place. The AI companies for example just pirate everything from Anna’s Archive. And they’re rich enough to afford enough lawyers to get away with that. And that’s unlike libraries, which pay for books and DVDs in their shelves… So that’s definitely illegal by any standard.
You can make temporary copies of copyrighted materials for fair use applications. I seriously hope there isn’t a state out there that is going to pass laws that gut the core freedoms of art, research, and basic functionality of the internet and computers. If you ban temporary copies like cache, you ban the entire web and likely computers generally, but you never know these days.
Know your rights and don’t be so quick to bandwagon. Consider the motives behind what is being said, especially when it’s two entities like these battling it out.
You have to remember, AI training isn’t only for mega-corporations. By setting up barriers that only benefit the ultra-wealthy, you’re handing corporations a monopoly of a public technology by making it prohibitively expensive to for regular people to keep up. These companies already own huge datasets and have whatever money they need to buy more. And that’s before they bind users to predatory ToS allowing them exclusive access to user data, effectively selling our own data back to us. What some people want would mean the end of open access to competitive, corporate-independent tools and would leave us all worse off and with fewer rights than where we started.
The same people who abuse DMCA takedown requests for their chilling effects on fair use content now need your help to do the same thing to open source AI. Their next greatest foe after libraries, students, researchers, and the public domain. Don’t help them do it.
I recommend reading this article by Cory Doctorow, and this open letter by Katherine Klosek, the director of information policy and federal relations at the Association of Research Libraries. I’d like to hear your thoughts.
Fuck that guy.
Fair use isn’t a loophole, it is copyright law.
I think it’s really disingenuous to mention the DeviantArt/Midjourney/Runway AI/Stability AI lawsuit without talking about how most of the infringement claims were dismissed by the judge.
Damn, this article is so biased.
Have you read this article by Cory Doctorow yet?
It’s local. You’re not sending data to their servers.
We’re looking at how we can use local, on-device AI models – i.e., more private – to enhance your browsing experience further. One feature we’re starting with next quarter is AI-generated alt-text for images inserted into PDFs, which makes it more accessible to visually impaired users and people with learning disabilities. The alt text is then processed on your device and saved locally instead of cloud services, ensuring that enhancements like these are done with your privacy in mind.
At least use the whole quote.
Have you heard of Fightcade?
Have you tried Bing Chat? It uses GPT 4 and has access to the internet. I ran your question through and it gave me this.
Oh, my bad.
You’re allowed to train on copyrighted works, it isn’t illegal for anybody. This article by Kit Walsh does a good job of breaking it down. She’s a senior staff attorney at the EFF.
Record them anyway. There’ll be more ways to de-anonymize them in the future.