Reddit CEO Steve Huffman is standing by Reddit’s decision to block companies from scraping the site without an AI agreement.
Last week, 404 Media noticed that search engines that weren’t Google were no longer listing recent Reddit posts in results. This was because Reddit updated its Robots Exclusion Protocol (txt file) to block bots from scraping the site. The file reads: “Reddit believes in an open Internet, but not the misuse of public content.” Since the news broke, OpenAI announced SearchGPT, which can show recent Reddit results.
The change came a year after Reddit began its efforts to stop free scraping, which Huffman initially framed as an attempt to stop AI companies from making money off of Reddit content for free. This endeavor also led Reddit to begin charging for API access (the high pricing led to many third-party Reddit apps closing).
In an interview with The Verge today, Huffman stood by the changes that led to Google temporarily being the only search engine able to show recent discussions from Reddit. Reddit and Google signed an AI training deal in February said to be worth $60 million a year. It’s unclear how much Reddit’s OpenAI deal is worth.
Huffman said:
Without these agreements, we don’t have any say or knowledge of how our data is displayed and what it’s used for, which has put us in a position now of blocking folks who haven’t been willing to come to terms with how we’d like our data to be used or not used.
“[It’s been] a real pain in the ass to block these companies,” Huffman told The Verge.
Honestly, any platforms hosting user-generated content who use the legal argument that they only provide hosting and aren’t responsible for what their user post shouldn’t also be able to sell the same data and claim owning any of it.
Otherwise, take away their legal immunity. Nazis or pedophiles post something awful? You get in front of the judge.
edit: typo
Exactly this. You can claim that their scraping is abusing your servers, but the moment you claim copyright for the content of the site, then you give up your Section 230 rights.
You’d also probably lose a whole lot more processing power trying to stop the crawlers vs just letting them have API access with some sort of limit to queries.
Eh, not really.
I block bot user agents to my Lemmy instance, and the overhead is pretty negligible for that (it’s all handled in my web firewall/load balancer).
Granted, those are bots that correctly identify themselves via user agent and don’t spoof a browser’s.
It’s also cheaper and easier to add another load balancer than size up or scale out my DB server to handle the bot traffic.
I don’t think they actually block malicious bots, the change they’ve made is just to the robots.txt, they don’t have to do anything.
Robots.txt does literally nothing. It’s a piece of courtesy that’s easily ignored if you don’t care.
Yeah but it stops bing and a bunch of AI scrapers that want to act like they’re following the rules
How do we know it stops bing? As far as anyone knows they could have instructed their programmers to alter the crawlers so that it ignores robots.txt when on Reddit - that should have taken them a whole 2 minutes.
Reddit blocking any search crawler via robots.txt is such a non thing that it shouldn’t even be reported.
Ok but they do respect it, we know that https://searchengineland.com/microsoft-confirms-reddit-blocked-bing-search-444385
They even have a page telling you how to use it https://blogs.bing.com/webmaster/June-2008/Robots-Exclusion-Protocol-joining-together-to-pro/
Can’t sell something you don’t own.
So if they’re selling the parts people want, they need to own the parts no one wants.
Well, you can give money to Reddit for a piece of paper, but unless Reddit is claiming copyright to the content posted there, then they can’t sue anyone for not paying. It would be very interesting to see the text of these “licensing agreements”.
They’re not claiming copyright. They have a perpetual, non-revokable license to the content, granted by the people who use their site when they post the content.
Good point!
Robots.txt isn’t a binding agreement, this isn’t stopping anyone for whom their drive for profit outweighs their ethics.
Also, Fuck Spez.
Honestly, my biggest issue with LLMs is how they source their training data to create “their own” stuff. A meme calling it a plagiarism machine struck a chord with me. Almost anyone else I’d sympathize with, but fuck Spez.
What resonated with me is people calling LLMs and Stable Diffusion “copyright laundering”. If copyright ever swung in AI’s favor it would be super easy to train an AI on stuff you want to steal, add in some generic training, and now you have a “new” piece of art.
LLMs and Stable Diffusion are just compression algorithms for abstract patterns, only one level above data.
The real takeaway of all of this is that copyright law is massively out of date and not fit for purpose in the 21st century or frankly the late 20th.
The current state of copyright law cannot deal with the internet, let alone AI
Yep they now get paid for the data we have them. I have no sympathy lol. At least these models can’t actually store it all losslessly by any stretch of the imagination. The compression factors would have to be like 100-200X+ anything we’ve ever been able to achieve before. The numbers don’t work out. The models do encode a lot though and some of it is going to include actual full text data etc but it’ll still be kinda fuzzy.
I think we do need ALL OPEN SOURCE. Not just for AI, but I know on that point I’m preaching to the choir here lol
The enshittification cycle:
Phase one, attract users by providing a good service.
Phase two, once the users are locked in, squeeze them for all they’re worth by selling them to business customers (advertisers and/or data buyers).
Phase three, once the business customers are locked in, squeeze them for all they’re worth by threatening to deny them access to the users on whom they now depend.Spez seems to think Reddit has the pull to make phase 3 happen. I rather doubt it, but we’ll see.
https://catvalente.substack.com/p/stop-talking-to-each-other-and-start
Blog post (?) from Catherynne Valente about this exact topic
Stop benefitting from the internet, it’s not for you to enjoy, it’s for us to use to extract money from you. Stop finding beauty and connection in the world, loneliness is more profitable and easier to control.
Spez is tracing Elon’s steps with X really closely.
Also fuck spez.
I’m sure plenty of others join me in the sentiment of thinking “Who the fuck are you to restrict MY free content that I contributed?”
God, fuck reddit so fucking hard
Fuck Spez. He’s probably editing the comments anyway, he literally can’t help himself.
profiting off of user generated content 😒
And yet reddit is happy to make money off our content for free.
Or at least it did. Personally I overwrote and deleted all my content a while back.
You think that Reddit didn’t already have the previous content saved?
Bingo, the only winning move is not to play at all and stop using Reddit.
Everyone always says this like it’s some kind of gotcha, but all of my nuked posts still have my “fuck you, reddit” content and haven’t been reverted. It’s been nearly exactly a year.
Maybe reddit has an offline copy of my old content and that of others somewhere, but if so they’d be handing that directly over to whoever under some kind of agreement – that certainly wouldn’t be the subject of any kind of site crawling which is the crux of the issue here.
You’re ignoring the idea that they could still be working on a way to restore content and haven’t completed that process yet
Or that they could start feeding your archived (not cached) data directly to the AI companies anyway for a price
IMO, you can win by jamming your “transmissions” with noise. It’s easier to hide in noise as noise than it Is to be silent IMO. Muddy the waters as it were
You’re ignoring the idea that they could still be working on a way to restore content and haven’t completed that process yet
there’s no evidence to suggest this, though.
Content is absolutely archived and they have financial incentive to restore the quality of their “knowledge base”
That’s a fair amount of circumstance and motivation to support my idea, regardless of tangible evidence
Motivation and circumstance, absent actual evidence, does not make for a convincing argument.
Alright well I guess evidence is needed before we can have ideas - crazy
i went looking for old comments and posts i had made after i overwrote then wiped them. They’re still gone. i looked again several months later, and they were still gone.
so, unless reddit did a massive restore of everyone’s comments/posts except for my 4 accounts, then i don’t believe they did it at all except for a select number of top contributors who deleted their content.
On the other side of the same coin: When I mass edited my comments before quitting Reddit, I got site-banned. Basically, my first account’s automated edit got me auto-banned from several subs with pro-spez mods. Some subs had set their automod to detect when people were using the more popular methods of auto-editing, and set the automod to ban for using them. Then when I did the same with my second (and third, and fourth, and fifth, etc…) account, it almost immediately got site-banned for ban evasion.
Basically, account 1 was banned from a sub, so when account 2 started doing the same thing on the same IP address, it was flagged as ban evasion. And ban evasion is one of the few things that will get you banned site-wide instead of just from a specific sub.
I went back and checked a few months ago, and all of those site bans were lifted and the edits were undone. Likely because a site ban prevents the comments from showing up (which hurts Reddit’s bottom line, because they show up as a bunch of [removed] comments instead,) but also prevented any of the edits from actually being published. So when they lifted the site ban (to get those old comments to show back up again) it was as if I had never edited them at all. I had probably a million karma spread across my various accounts. I was extremely active at one point, so Reddit had a direct incentive to unban those accounts with literal thousands of comments.
As I said in another comment, I was not suggesting that Reddit would restore your comments to public view.
there’s no evidence to suggest that, either.
Except for incidents like This
There’s no evidence of the contrary either.
You are assuming edits overwrite existing content. Instead of overwriting, they could just store the edited post as a new entry in the database with a higher version number. Then, you only show the latest version of each post to the end users while keeping the older versions available die Reddit’s own use.
In fact, it is extremely likely they do this. It is basically a necessity if you want to be able to properly moderate a site like Reddit. Otherwise you could simply post spam or unsavory content, and then overwrite it with something benign an hour or so later, before there were enough reports and a moderator would have gotten a chance to review it.
You are assuming edits overwrite existing content
i have seen no evidence to suggest otherwise, but thanks for sharing your theories
In fact, it is extremely likely they do this
based on what evidence? your baseless speculation?
The fact that they managed to restore overwritten posts after people started to delete their history.
this could also be explained by sketchy scripts failing to completely delete posts/comments, which i even noticed myself when checking that they had done their jobs properly. as i mentioned in another comment, i had to run the shredder scripts several times for complete overwrite/deletion. or it could be database errors failing to register edits/deletions due to extremely heavy loads at the time. it could be a lot of things.
the point is that we don’t have any direct evidence of what it actually was, just a lot of circumstantial evidence and a lot of speculation. nothing definitive.
Reddit used to be open source. There is still a copy of that source available on github. It’s 7 years old so it’s probably significantly different from what they are running now. Still, it gives some insight into the design.
For example, deleted comments aren’t deleted, it just sets a deleted flag. Example code that shows this.
I haven’t dug around the code enough to figure out how editing works, it’s Python code so an unreadable mess. The database design also seems very strange. It’s like they built a database system on top of a database.
Reddit is dying anyway.
It’s easy to say this as someone who is “on the other side”. But the data doesn’t really back up that statement.
What if I had an agreement with MS that they can scrape my data and anything I post online?
If you’re still donating your content to the Facebook, Twitter and Reddit data stores, then I don’t know what to tell you man, you’re literally enriching the worst people, who will do the worst things, with your information, your stories, your answers, your comments, your labor, your effort. You are giving yourselves to them. Literally.
Fuck Reddit.
This deal is ridiculous and a terrible precedent for the Internet moving forward. Imagine having to juggle multiple search engines in order to actually search the Internet for things you were looking for because the results literally can’t show up.
What happens when certain news organizations only strike a deal with Bing? And certain forums only strike to deal with Google? This is so shortsighted and they didn’t even get an appreciable amount of money for fucking us over.