We develop data sets to train our algorithms so that we can improve the services we provide to customers like you. We have devoted significant time and resources to developing methods to ensure that these data sets are anonymized and de-identified.
To develop these data sets, we sample snippets of text at random, disassociate them from a user's account, and then use a variety of different methods to strip the text of identifying information (such as identifiers, contact details, addresses, etc.). Only then do we use the snippets to train our algorithms-and the original text is deleted. In other words, we don't store any text in a manner that can be associated with your account or used to identify you or anyone else.
We currently offer a feature that permits customers to opt out of this use for Grammarly Business teams of 500 users or more. Please let me know if you might be interested in a license of this size, and I'II forward your request to the corresponding team.
Per their website premium includes "Unlimited sentence paraphrasing powered by A.I." so I'm not sure they're an appropriate alternative to avoid the "AI" bullshit.
I'm pretty sure most tools like this have to use ai to some degree to be more effective than something like Microsoft Word. I think the issue is more whether it's opt in or not to include your own data.
I have. It's pretty short and to the point. They're based out of Germany so their requirements for clarity are pretty high by law. They go into quite a lot of detail about what is sent.
In this case they send date, time, language, processing time, number and the type of errors, but not the text itself
However, they do have an optional feature that uses OpenAI to rephrase sentences so that might be training through the back door.
I've been using it for years and have been very happy with the service.
I took a quick look at this and it seems that the server portion of this product is open source but the apps such as extensions are not. I'm not saying it's bad or even that it's a red flag. I just felt like I should point it out.
True. Companies sell our data to third parties since forever, but some people are worried about it being used to train machine learning models? I'm far more concerned by people using it than AI.
It's because certain companies are stirring the pot and manipulating. They want people mad so they can put restrictions on training AI, to stifle the open source scene.
I see you posted this article to 4 communities. According to the comments on this post if you use the cross post function (in the default web frontend), it will only show once in the feeds instead of 4 times (which can be a bit annoying).
I did use the cross-post function. Most apps do not currently acknowledge this function which might explain why the article has appeared to you multiple times.
I see content from many servers in the lemmy federation. My understanding, which could be wrong, was that like email, you can post to any domain and see posts from other domains. What's the advantage of posting to many instances?
Even as someone who declines all cookies where possible on every site, I have to ask. How do you think they are going to be able to improve their language based services without using language learning models or other algorithmic evaluation of user data?
I get that the combo of AI and privacy have huge consequences, and that grammarly's opt-out limits are genuinely shit. But it seems like everyone is so scared of the concept of AI that we're harming research on tools that can help us while the tools which hurt us are developed with no consequence, because they don't bother with any transparency or announcement.
Not that I'm any fan of grammarly, I don't use it. I think that might be self-evident though.
Framing this solely as fear is extremely disingenuous. Speaking only for myself: I'm not against the development of AI or LLMs in general. I'm against the trained models being used for profit with no credit or cut given to the humans who trained it, willing or unwilling.
It's not even a matter of "if you aren't the paying customer, you're the product" - massive swaths of text used to train AIs were scraped without permission from sources whose platforms never sought to profit from users' submissions, like AO3. Until this is righted (which is likely never, I admit, because the LLM owners have no incentive whatsoever to change this behavior), I refuse to work with any site that intends to use my work to train LLMs.
I'm against the trained models being used for profit with no credit or cut given to the humans who trained it.
Sorry mate, hell's gonna get cold before this happens. We're talking about the biggest moth******ers on earth since always. Do you think Meta/[insert big tech company name here] will start to behave all of the sudden? These people literally KILL people everyday for a profit (looking at you Instagram).
The only way to get something from these scumbags is fining them something like 100k per hour, until they start respecting people's privacy
Models need vast amounts of data. Paying individual users isnt feasible, and like you said most of it can be scraped.
The only way I see this working is if scraped content is a no go and then you pay the website, publishing house, record company, etc which kills any open source solution and doesn't really help any of the users or creators that much. It also paves the way for certain companies owning a lot of our economy as we move towards an AI driven society.
It's definitely a hot mess but the way I see it, the more restrictive we are with it, the more gross monopolies we create for no real gains.
They're honestly doing you a favor. Grammarly is terrible. I've seen some of my friends whose first language isn't English use it to try to clean their grammar up and it makes some really weird, often totally mistaken choices. Usually they would have been better off leaving it as they wrote it.
They have a free tier and a $10/mo tier and prominently advertise their AI without any information about privacy. Guaranteed you and your text are the product being used to train their AI.
Think about this every time you or a project you contribute to is using Microsoft GitHub instead of an open source offering (or self-hosted) or folks contributing to your permissive-licensed project living elsewhere while using Microsoft GitHub Copilot. All your projects and that force-push history clean up now belong to the Microsoft-owned AI that sells itself back to the developers that wrote all the code it trained on—no compensation, no recognition.
I am perfectly fine with providing training data for AI, and have actually spent hours contributing to various projects. However, it is super scummy for a company to collect and use sensitive user data (literally a keylogger) not only without any form of communication or consent, but where the only way to opt out is to pay.
Uhh umm... You are the product! Aaand... Shill for greedy corporations!
I remember when Google said quite openly that they'd give us email addresses with more storage than we'd ever dreamed for life and in return, they'd scan the first few sentences of all messages and use them to target ads at us and we were all like, "Sounds fair."
Any scope of privacy conscious users banding together to create a shell corp to pay for a business account? 500 users sounds doable. More the merrier, yeah?