ℍ𝕂-𝟞𝟝

  • 2 Posts
  • 1.1K Comments
Joined 2 years ago
cake
Cake day: July 14th, 2024

help-circle


  • CVNs and a larger navy are only useful if we want to get into the power projection game, and even then, we have some stuff already. I’m sure we could still do something like the Falklands war. The advamtage the US has its also its disadvantage, as we don’t need a fleet to get to the Middle East or Africa.

    We have SSBNs enough that makes attacking the EU idiotic for even the US.

    We do need strategic stuff like AWACS planes and more satellites though, but what we need most is standardisation, since there are like 11 MBTs in service in the EU right now.



  • Yeah but it doesn’t matter what the objective of the scraper is, the only thing that matters is that it’s an automated client that is going to send mass requests to you. If it wasn’t, Anubis would not be a problem for it.

    The effect is the same, increased hosting costs and less access for legitimate clients. And sites want to defend against it.

    That said, it is not mandatory, you can avoid using Anubis as a host. Nobody is forcing you to use it. And as someone who regularly gets locked out of services because I use a VPN, Anubis is one of the least intrusive protection methods out there.





  • AI does not triple traffic. It’s a completely irrational statement to make.

    Multiple testimonials from people who host sites say they do. Multiple Lemmy instances also supported this claim.

    I would bet that the number of requests per year of s resource by an AI scrapper is on the dozens at most.

    You obviously don’t know much about hosting a public server. Try dozens per second.

    There is a booming startup industry all over the world training AI, and scraping data to sell to companies training AI. It’s not just Microsoft, Facebook and Twitter doing it, but also Chinese companies trying to compete. Also companies not developing public models, but models for internal use. They all use public cloud IPs, so the traffic is coming from all over incessantly.

    Using as much energy as a available per scrapping doesn’t even make physical sense. What does that sentence even mean?

    It means that Microsoft buys a server for scraping, they are going to be running it 24/7, with the CPU/network maxed out, maximum power use, to get as much data as they can. If the server can scrape 100 sites per minute, it will scrape 100 sites. If it can scrape 1000, it will scrape 1000, and if it can do 10, it will do 10.

    It will not stop scraping ever, as it is the equivalent of shutting down a production line. Everyone always uses their scrapers as much as they can. Ironically, increasing the cost of scraping would result in less energy consumed in total, since it would force companies to work more “smart” and less “hard” at scraping and training AI.

    Oh, and it’s S-C-R-A-P-I-N-G, not scrapping. It comes from the word “scrape”, meaning to remove the surface from an object using a sharp instrument, not “scrap”, which means to take something apart for its components.



  • Websites were under a constant noise of malicious requests even before AI, but now AI scraping of Lemmy instances usually triples traffic. While some sites can cope with this, this means a three-fold increase in hosting costs in order to essentially fuel investment portfolios.

    AI scrapers will already use as much energy as available, so making them use more per site measn less sites being scraped, not more total energy used.

    And this is not DDoS, the objective of scrapers is to get the data, not bring the site down, so while the server must reply to all requests, the clients can’t get the data out without doing more work than the server.