I've said so many wrong and stupid things on reddit over the years both purposely and not. I wouldn't trust anything trained on reddit. And facebook shouldn't even be thought if as a training source as its mainly bots, idiots and propaganda.
Stack overflow, I've never used but seems to be mostly smart people.
Cool! So they're training AI on the worst humanity has to offer: Young Nazis for Freedom, "No Means Try Again Later" Techbrodudes, and Old Nazis who Don't Care About Freedom.
If they hadn't made it impossible to use a functional app to browse their site, I probably wouldn't care.
But the fact that they blocked 100 apps that would, if you merged the absolute worst of all of them, still blow the actual Reddit app out of the water? Nope. Fuck 'em.
According to gdpr it is not allowed to use people's data for purposes other than the ones they agreed to.
I had an e-mail discussion with them and I filed a complaint with my countries gdpr enforcement agency.
They take a while to investigate and react, but hopefully if enough people complain they will take it seriously.
An AI using what I say to train it will only make AI more like me. As far as I'm concerned, that's an improvement that may help others while not affecting my life one iota. It's not like it remembers who said what, it all just tweaks the algorithm.
Advertisers have been recording and storing what you say word for word to build a profile specific to you, containing your deepest, darkest truths so they can trick you into giving them more of your money. Who cares about AI training on publicly posted comments, it's just taking energy away from the real privacy issues.
In 2022 I was among the 1% of most karma awarded on reddit. Hadn't been banned anywhere.
Then in 2024 their IPO went public. A week later it was said that AI bots were now scrubbing the site to sell your data to google.
Within 3 weeks I was permanently banned, for 3 posts that had nothing wrong with them. One of them I understand how an AI would flag if its just using keywords. In context of the conversation nothing was wrong.
The other 2 didn't even have keywords which an AI would pick up on.
So my assumption is the AI scrubbed my data, started acting like me, which is to say an absurdist, and google didn't like the results of my data. So reddit banned me.
Exactly!!! All this fear about AI being trained as if our data hasn't been being vacuumed for years and years now. Tracking profiles, fingerprinting, etc. There are legitimate things to get energized and concerned over, no need to make up fear-based narratives
You retain any ownership rights you have in Your Content, but you grant Reddit the following license to use that Content:
When Your Content is created with or submitted to the Services, you grant us a worldwide, royalty-free, perpetual, irrevocable, non-exclusive, transferable, and sublicensable license to use, copy, modify, adapt, prepare derivative works of, distribute, store, perform, and display Your Content and any name, username, voice, or likeness provided in connection with Your Content in all media formats and channels now known or later developed anywhere in the world. This license includes the right for us to make Your Content available for syndication, broadcast, distribution, or publication by other companies, organizations, or individuals who partner with Reddit. You also agree that we may remove metadata associated with Your Content, and you irrevocably waive any claims and assertions of moral rights or attribution with respect to Your Content.
Seriously. Eating Reddit means some truly strange shit is going to emerge. We all know the pattern of fun replies on Reddit, but the AI does not. Hence the glue thing.
I don't think that's an impossible problem. Existing models can reliability distinguish between, for example, different languages. Most of their training data is presumably in English but while this may make them better at generating English text, it doesn't make them randomly switch from other languages to English. A sufficiently advanced model would likewise distinguish between descriptions of reality and shit-posts because the content of shit-posts would not be useful for predicting descriptions of reality. Some fine tuning would teach it to produce just the descriptions of reality.
Or look at it this way: the folks developing these LLMs aren't ignorant of the fact that Reddit content is often false and meant to be funny. They're not going to make the sort of silly mistake that someone who isn't an expert can still easily predict and they're not going to train their LLMs on that content if it makes the LLMs worse, although we're still going to see some glue on pizza while the technology continues to develop.
It can cross check a language with tons of other words and examples of that language already in its data set. There is no such data for whether or not something confirms with reality. That simply doesn't exist and really won't ever exist. They are not similar problems. One is immensely more challenging to solve than the other.
In my opinion stack overflow has vague answers 80% of the time. If AI helps filtering all the bullshit, I’m ok with that. And I don’t know how Reddit or Facebook bot comments are a good source of information.
They are using your knowledge and talents to profit immensely by selling or using it. Selling it right back to you and everyone else.
I have a problem with that.
Consider the source of the Reddit comments, Stack Overflow questions and FB photos: a substantial number of them are poorly written, irrelevant, re-posts and were already churned out by AI. So the ouroboros is the master of and source of its own fate. GIGO.
So fill yer boots, AI, it's your own shit you're eating.