• nialv7@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    6 days ago

    We had a trust based system for so long. No one is forced to honor robots.txt, but most big players did. Almost restores my faith in humanity a little bit. And then AI companies came and destroyed everything. This is why we can’t have nice things.

  • mfed1122@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    7 days ago

    Okay what about…what about uhhh… Static site builders that render the whole page out as an image map, making it visible for humans but useless for crawlers 🤔🤔🤔

      • mfed1122@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        1
        ·
        7 days ago

        I wasn’t being totally serious, but also, I do think that while accessibility concerns come from a good place, there is some practical limitation that must be accepted when building fringe and counter-cultural things. Like, my hidden rebel base can’t have a wheelchair accessible ramp at the entrance, because then my base isn’t hidden anymore. It sucks that some solutions can’t work for everyone, but if we just throw them out because it won’t work for 5% of people, we end up with nothing. I’d rather have a solution that works for 95% of people than no solution at all. I’m not saying that people who use screen readers are second-class citizens. If crawlers were vision-based then I might suggest matching text to background colors so that only screen readers work to understand the site. Because something that works for 5% of people is also better than no solution at all. We need to tolerate having imperfect first attempts and understand that more sophisticated infrastructure comes later.

        But yes my image map idea is pretty much a joke nonetheless

  • SufferingSteve@feddit.nu
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    8 days ago

    There once was a dream of the semantic web, also known as web2. The semantic web could have enabled easy to ingest information of webpages, removing soo much of the computation required to get the information. Thus preventing much of the AI crawling cpu overhead.

    What we got as web2 instead was social media. Destroying facts and making people depressed at a newer before seen rate.

    Web3 was about enabling us to securely transfer value between people digitally and without middlemen.

    What crypto gave us was fraud, expensive jpgs and scams. The term web is now even so eroded that it has lost much of its meaning. The information age gave way for the misinformation age, where everything is fake.

      • Serinus@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        7 days ago

        I feel like half of the blame capitalism gets is valid, but the other half is just society. I don’t care what kind of system you’re under, you’re going to have to deal with other people.

        Oh, and if you try the system where you don’t have to deal with people, that just means other people end up handling you.

        • kazerniel@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          7 days ago

          It matters a lot though what kind of goal the system incentivises. Imagine if it was people’s happiness and freedom instead of quarterly profits.

        • Amju Wolf@pawb.social
          link
          fedilink
          English
          arrow-up
          1
          ·
          7 days ago

          In this case it is purely fault of the money incentive though. Noone would spend so much effort and computation power on AI if they didn’t think it could make them money.

          The funniest part is though that it’s only theoretical anyway, everyone is only losing on it and they’re most likely never gonna make it back.

  • zbyte64@awful.systems
    link
    fedilink
    English
    arrow-up
    1
    ·
    7 days ago

    Is there nightshade but for text and code? Maybe my source headers should include a bunch of special characters that then give a prompt injection. And sprinkle some nonsensical code comments before the real code comment.

  • zifk@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 days ago

    Anubis isn’t supposed to be hard to avoid, but expensive to avoid. Not really surprised that a big company might be willing to throw a bunch of cash at it.

    • BlameTheAntifa@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 days ago

      The problem is that hundreds of bad actors doing the same thing independently of one another means it does not qualify as a DDoS attack. Maybe it’s time we start legally restricting bots and crawlers, though.

  • Spaz@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    6 days ago

    Is there a migration tool? If not would be awesome to migrate everything including issues and stuff. Bet even more people would move.

    • BlameTheAntifa@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 days ago

      Codeberg has very good migration tools built in. You need to do one repo at a time, but it can move issues, releases, and everything.

  • PhilipTheBucket@piefed.social
    link
    fedilink
    English
    arrow-up
    0
    ·
    8 days ago

    I feel like at some point it needs to be active response. Phase 1 is a teergrube type of slowness to muck up the crawlers, with warnings in the headers and response body, and then phase 2 is a DDOS in response or maybe just a drone strike and cut out the middleman. Once you’ve actively evading Anubis, fuckin’ game on.

    • TurboWafflz@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 days ago

      I think the best thing to do is to not block them when they’re detected but poison them instead. Feed them tons of text generated by tiny old language models, it’s harder to detect and also messes up their training and makes the models less reliable. Of course you would want to do that on a separate server so it doesn’t slow down real users, but you probably don’t need much power since the scrapers probably don’t really care about the speed