this post was submitted on 09 Jun 2023
466 points (98.7% liked)

Memes

45740 readers
1456 users here now

Rules:

  1. Be civil and nice.
  2. Try not to excessively repost, as a rule of thumb, wait at least 2 months to do it if you have to.

founded 5 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] sealneaward@lemmy.ml 2 points 1 year ago (1 children)

Creating a web scraper vs actually maintaining one that is effective and works is two different things. It's very easy to fight web scraping if you know what you are doing.

[–] argv_minus_one@beehaw.org 1 points 1 year ago* (last edited 1 year ago) (1 children)

Right, but these are big companies with lots of talented programmers on hand. If anyone can overcome such an obstacle, it's them.

Also, Google and Microsoft already have a search index full of Reddit content to scrape.

[–] sealneaward@lemmy.ml 1 points 1 year ago (2 children)

You are right. You would need a team of skilled scrapers and network engineers though would know how to get around rate limiters with some kind of external load balancer or something along those lines.

[–] MrPoopyButthole@lemmy.world 2 points 1 year ago

Rate limiters work on IP source. This is easily bypassed with a rotating proxy. There are even SaaS that offer this. The trick is to not use large subnets that can be easily blocked. You have to use a lot of random /32 IPs to be effective.

[–] argv_minus_one@beehaw.org 1 points 1 year ago (1 children)

That problem is already solved. Google and Microsoft are already fetching every single page on Reddit for search engine indexing.

[–] sealneaward@lemmy.ml 1 points 1 year ago (1 children)

Could they be doing that already because of the still open API of Reddit and that will soon change? I just feel like it's easier for them currently and it will be tougher once the API changes are implemented.

[–] argv_minus_one@beehaw.org 2 points 1 year ago

No. Search engines fetch pages using plain old HTTP GET requests, same as how browsers fetch pages. There is some difficulty in parsing the HTML and extracting meaningful content, but it's too late: the HTML is already stored on Google/Microsoft servers, ready for extraction, and there's nothing Reddit can do to stop them.

Reddit can make future content harder to extract, but not without also making it invisible to search engines, which would cause Reddit to disappear from Google Search and Bing.

That's why I say trying to charge money for AI training data is a fool's errand. These facts make it impossible. That doesn't mean Spez won't try, but it does mean he won't succeed.