Hacker Newsnew | past | comments | ask | show | jobs | submit | xiamx's commentslogin

Same. And opening the page and seeing that it talks about myopia is a surprise


Reminds me of a (personal) website I saw long time ago that replicates an earlier day OSX desktop, with functional itunes player. Does anyone remember what it was called?


I don't recall that exact one but I do love personal websites designed as OS's, not sure why, just my thing.


Try your technique on a few of these fingerprint testing sites https://github.com/niespodd/browser-fingerprinting#fingerpri... I'm pretty sure it's quite detectible


Hmm maybe I will if I have time. We've been using this technique for user-initiated scraping. The only issue we've run in to is we get rate-limited by IP sometimes. Changing the IP has solved the problem each time.


If I am correct in assuming the parent is talking about puppeteer, there is a plugin[1] that claims to evade most of the methods used to detect headless browsers. I have used it recently for just that purpose, and I can say that it worked wifh minimal setup and configuration for my usecase, but I guess depending on the detection mechanisms youre evading YMMV.

The creator of that plugin does mention it is very much a cat and mouse game, just like most of the “scraping industry”

https://www.npmjs.com/package/puppeteer-extra-plugin-stealth


I think it would've been even more immersive if the screen was warped to cover the CRT


On that theme, have you tried cool-retro-term?

It's nostalgic.


It's the same person that wrote bpytop. He also wrote bashtop. 3 times implementation of the same thing. That's dedication!


Yeah, a blog post like this is too casual and generalizes way too much


On appstore tweetbot does have a "for Twitter" suffix to its name though


Why would they do that. These guys are generating revenue by offering managed service, not by selling opensource software.


For the same reasons MongoDB and Redis changed their licenses to avoid AWS et al offering it as a managed service?


That's a different story. Aws was competiting for revenue with the developers of MongoDB and Redis' for profit company. Postgres' publisher does not use managed service as a source of income.


Of course, but I was just offering a reason for why they would want to do that. It’s not as if it’s unreasonable if they did that.


What's wrong with scraping a bunch of pages. As long as they are following robots.txt, it's no big deal.


Or even better, have contracts with the companies. Maybe unlikely for them, but I think “scraping” is too often assumed to be “bad” in some way. The company I work for does a lot of web scraping, but we have contracts with our partners to scrape their websites. They may still have robots.txt that ask users not to scrape some areas, but we are allowed to bypass those.


HN always had a boner for web-crawler hate. How dare you automate downloading of public data others post online!


So proofhub names itself the first in its list, sounds legit


Many have done the same.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: