1. Google’s “only‑scraper” narrative
Many commenters see Google’s move to block the IPIDEA proxy network as a way to keep competitors out.
“Only Google is allowed to scrape the web.” – a456463
“Google must be the only one allowed to scrape the web.” – a456463
“Google is trying to pull the ladder up behind them and make it more difficult for other companies to collect training data.” – a456463
2. Residential proxies – double‑edged sword
The discussion splits between legitimate use (data‑collection, research) and abuse (spam, botnets).
“Residential proxies are the only way to crawl and scrape.” – tonymet
“They are good. Ones which you pay for and which are running legitimately, with the knowledge (and compensation) of those who run them.” – progbits
“Malware in random apps running on your device without your knowledge is bad.” – throwoutway
3. Site‑owner pain from proxy‑driven traffic
Operators complain that residential‑proxy traffic overwhelms their defenses and forces stricter blocking.
“The residential proxies are a real pain. Tons and tons of abusive traffic.” – the_fall
“I had to tighten my Cloudflare WAF rules a lot.” – Kodiack
“Excessive traffic from residential proxies is a pain for site operators.” – Kodiack
4. Scraping etiquette and anti‑scraping measures
The debate centers on whether scraping is legal, how to respect robots.txt, and what tools (rate‑limits, IP ranges) should be used.
“Scraping is a perfectly legal activity, after all.” – ronsenshi
“Google allows access to GoogleBot but not others.” – ronsenshi
“Scraping is legal but problematic; need for robots.txt, rate limits.” – toofy
These four themes capture the main currents of opinion in the thread.