- 0 Posts
- 101 Comments
sudo@programming.devto
Today I Learned@lemmy.world•TIL that trolls working for the Russian Government helped divide the american feminist movementEnglish
243·14 days agoLike the NYT suddenly cares about Palestinian women.
For more than a century, Russia and the Soviet Union sought to weaken their adversaries in the West by inflaming racial and ethnic tensions.
FUCK ALL THE WAY OFF, NYT. DID YOU FORGET FRED HAMPTON? DID YOU FORGET THE FBI BLACKMAILING MLK?
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
3·14 days ago🤙
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
5·14 days agoAnubis forces the site to reload when doing the normal PoW challenge! Meta Refresh is a sufficient mouse to block 99% of all bot traffic without being any more burdensome than PoW.
You’ve failed to demonstrate why meta-refresh is more burdensome than PoW and have pivoted to arguing the point I was making from the start as though it was your own. I’m not arguing with you any further. I’m satisfied that I’ve convinced any readers of our discussion.
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
4·14 days agoYou will have people complain about their anti-fingerprinting being blocked with every bot-managment solution. Your ability to navigate the internet anonymously is directly correlated with a bots ability to scrape. That has never been my complaint about Anubis.
My complaint is that the calculations Anubis forces you to do are absolutely negligible burden for a bot to solve. The hardest part is just having a JavaScript interpreter available. Making the author of the scraper write custom code to deal with your website is the most effective way to prevent bots.
Think about how much computing power AI data centers have. Do you think they give a shit about hashing some values for Anubis? No. They burn more compute power than a thousand Anubis challenges generating a single llm answer. PoW is a backwards solution.
Please Think. Captchas worked because they’re supposed to be hard for a computer to solve but are easy for a human. PoW is the opposite.
In the current shape Anubis has zero impact on usability for 99% of the site visitors, not so with meta refresh.
Again, I ask you: What extra burden does meta-refresh impose on users? How does setting a cookie and immediately refreshing the page burden the user more than making them wait longer while draining their battery before doing the exact same thing? Its strictly less intrusive.
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
3·14 days agoAnd how do you actually check for working JS in a way that can’t be easily spoofed? Hint: PoW is a good way to do that.
Accessing the browsers API in any way is way harder to spoof than some hashing. I already suggested checking if the browser has graphics acceleration. That would filter out the vast majority of headless browsers too. PoW is just math and is easy to spoof without running any JavaScript. You can even do it faster than real JavaScript users something like Rust or C.
Meta refresh is a downgrade in usability for everyone but a tiny minority that has disabled JS.
What are you talking about? It just refreshes the page without doing any of the extra computation that PoW does. What extra burden does it put on users?
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
22·14 days agoLOL
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
31·14 days agoYou are arguing a strawman. Anubis works because because most AI scrapers (currently) don’t want to spend extra on running headless chromium
WTF, That’s what I already said? That was my entire point from the start!? You don’t need PoW to force headless usage. Any JavaScript challenge will suffice. I even said the Meta Refresh challenge Anubis provides is sufficient and explicitly recommended it.
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
41·14 days agoWell in most cases it would by Python requests not curl. But yes, forcing them to use a browser is the real cost. Not just in CPU time but in programmer labor. PoW is overkill for that though.
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
7·14 days agoAnubis is that it has a graded tier system of how sketchy a client is and changing the kind of challenge based on a a weighted priority system.
Last I checked that was just User-Agent regexes and IP lists. But that’s where Anubis should continue development, and hopefully they’ve improved since. Discerning real users from bots is how you do proper bot management. Not imposing a flat tax on all connections.
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
51·14 days agoThen there was a paper arguing that PoW can still work, as long as you scale the difficulty in such a way that a legit user
Telling a legit user from a fake user is the entire game. If you can do that you just block the fake user. Professional bot blockers like Cloudflare or Akamai have machine learning systems to analyze trends in network traffic and serve JS challenges to suspicious clients. Last I checked, all Anubis uses is User-Agent filters, which is extremely behind the curve. Bots are able to get down to faking TLS fingerprints and matching them with User-Agents.
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
52·14 days agoIts like you didn’t understand anything I said. Anubis does work. I said it works. But it works because most AI crawlers don’t have a headless browser to solve the PoW. To operate efficiently at the high volume required, they use raw http requests. The vast majority are probably using basic python
requestsmodule.You don’t need PoW to throttle general access to your site and that’s not the fundamental assumption of PoW. PoW assumes (incorrectly) that bots won’t pay the extra flops to scrape the website. But bots are paid to scape the website users aren’t. They’ll just scale horizontally and open more parallel connections. They have the money.
sudo@programming.devto
Selfhosted@lemmy.world•Anubis is awesome and I want to talk about itEnglish
568·14 days agoI’ve repeatedly stated this before: Proof of Work bot-management is only Proof of Javascript bot-management. It is nothing to a headless browser to by-pass. Proof of JavaScript does work and will stop the vast majority of bot traffic. That’s how Anubis actually works. You don’t need to punish actual users by abusing their CPU. POW is a far higher cost on your actual users than the bots.
Last I checked Anubis has an JavaScript-less strategy called “Meta Refresh”. It first serves you a blank HTML page with a
<meta>tag instructing the browser to refresh and load the real page. I highly advise using the Meta Refresh strategy. It should be the default.I’m glad someone is finally making an open source and self hostable bot management solution. And I don’t give a shit about the cat-girls, nor should you. But Techaro admitted they had little idea what they were doing when they started and went for the “nuclear option”. Fuck Proof of Work. It was a Dead On Arrival idea decades ago. Techaro should strip it from Anubis.
I haven’t caught up with what’s new with Anubis, but if they want to get stricter bot-management, they should check for actual graphics acceleration.
All of the AI uses I’ve listed have been around for almost a decade or more and are the only computational solutions to those problems. If you’ve ever used speech to text that wasn’t a speak-n-spell you were using a very basic AI model. If you ever scanned a document and had the text be recognized, that’s an AI model.
The catch here is I’m not talking about chatgpt or anything trying be very “general”. These are all highly specialized ai models that serve a very specific function.
The firefox AI sidebar embeds an external open-webui. It doesn’t roll its own ui for chat. Everything with AI is done in the quickest laziest way.
What exactly isn’t very open about open-webui or ollama? Are there some binary blobs or weird copyright licensing? What alternatives are you suggesting?
Basically everything its used for that isn’t being shoved in your face 24/7.
- speech to text
- image recognition
- image to text (includes OCR)
- language translation
- text to speech
- protein folding
- lots of other bio/chem problems
Lots of these existed before the AI hype to the point they’re taken for granted, but they are as much AI an LLM or image generator. All the consumer level AI services range from annoying to dangerous.
If I can pick my own API (including local) and sampling parameters
You can do this now:
- selfhost ollama.
- selfhost open-webui and point it to ollama
- enable local models in about:config
- select “local” instead of ChatGPT or w/e.
Hardest part is hosting open-webui because AFAIK it only ships as a docker image.
Edit: s/openai/open-webui
To build on this, it would help to install some sort of system monitoring to check temps, fanspeed, system usage and have those constantly going so OP can check for any red flags during a freeze.
Idk about that. In my case I believe my CPU was defective from the start and I lived with it because I always assumed it was my OS in some way.
If your CPU has seven years of not randomly freezing and its just doing that now then I wouldn’t suspect the CPU.
However, unless you find some clues from
journalctl -xeb1or dmesg I would assume its faulty hardware somewhere.



That seems like a flight of stairs up.