Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. The bright #LLM future, next part.

The bright #LLM future, next part.

Scheduled Pinned Locked Moved Uncategorized
llmgentoonoainollm
38 Posts 21 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

    The bright #LLM future, next part.

    git.gentoo.org is now effectively dead, being DDoS-ed by almost a million different IPs every day. Most of them are just performing a single request at a totally random URL. How are people supposed to deal with that? How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

    If you use LLM crap, you're part of the problem. You support these bastards. You should be ashamed of yourself.

    #Gentoo #NoAI #NoLLM #AI

    davidgerard@circumstances.runD This user is from outside of this forum
    davidgerard@circumstances.runD This user is from outside of this forum
    davidgerard@circumstances.run
    wrote last edited by
    #11

    @mgorny iocaine 3 works against this ok

    watch out for false positives

    villares@ciberlandia.ptV 1 Reply Last reply
    0
    • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

      The bright #LLM future, next part.

      git.gentoo.org is now effectively dead, being DDoS-ed by almost a million different IPs every day. Most of them are just performing a single request at a totally random URL. How are people supposed to deal with that? How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

      If you use LLM crap, you're part of the problem. You support these bastards. You should be ashamed of yourself.

      #Gentoo #NoAI #NoLLM #AI

      newcancerbero@masto.esN This user is from outside of this forum
      newcancerbero@masto.esN This user is from outside of this forum
      newcancerbero@masto.es
      wrote last edited by
      #12

      @mgorny Sorry but you are wrong. I use LLM and yesterday I give Importan information about attack of Mastodon social, and how was used that scum the A.I. Why? I know very much model, and I know than model her can use for a attack, I for my continue testing of A I. Can explain to moderador the nature of attack. Can you explain point by point was happened? I think the answer is NO. Think about this, you want fight blind, with clothes anti-bullet for stoping a Tank. Use your mind please.

      1 Reply Last reply
      0
      • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

        The bright #LLM future, next part.

        git.gentoo.org is now effectively dead, being DDoS-ed by almost a million different IPs every day. Most of them are just performing a single request at a totally random URL. How are people supposed to deal with that? How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

        If you use LLM crap, you're part of the problem. You support these bastards. You should be ashamed of yourself.

        #Gentoo #NoAI #NoLLM #AI

        algernon@come-from.mad-scientist.clubA This user is from outside of this forum
        algernon@come-from.mad-scientist.clubA This user is from outside of this forum
        algernon@come-from.mad-scientist.club
        wrote last edited by
        #13

        @mgorny

        How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

        Three ifs in a trenchcoat will get rid of the majority of those, without any additional software. The crawlers may appear complicated to defeat if you look at the user-agent only, but as soon as you look at some other headers, it turns out they're really, really, really dumb.

        If you want to do more than that, and do it slightly more efficiently than a reverse proxy can, iocaine can help.

        Unlike Anubis, the crawlers throwing more compute on it will not get past it, and legit visitors will (usually) remain unaware of its existence. It's in front of my own forge, happily serves ~800 req/sec (where the bottleneck is Caddy & TLS) on a €5/month potato quality VPS. It can also firewall IPs off, to further reduce load.

        It does catch some "legit" crawlers like Googlebot and Bingbot, but you can allow-list those, or keep them blocked because both of those feed into LLM training too.

        naturemc@mastodon.onlineN K cblgh@merveilles.townC 3 Replies Last reply
        0
        • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

          The bright #LLM future, next part.

          git.gentoo.org is now effectively dead, being DDoS-ed by almost a million different IPs every day. Most of them are just performing a single request at a totally random URL. How are people supposed to deal with that? How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

          If you use LLM crap, you're part of the problem. You support these bastards. You should be ashamed of yourself.

          #Gentoo #NoAI #NoLLM #AI

          kigelia@mastodon.onlineK This user is from outside of this forum
          kigelia@mastodon.onlineK This user is from outside of this forum
          kigelia@mastodon.online
          wrote last edited by
          #14

          @mgorny I am fighting a lone battle in my department at work against use of AI tools due to the environmental impact.

          However from someone nowhere near technical enough to understand this completely. Is this post saying essentially that the crawling of the internet for input to ‘learn’ from by AI companies is clogging up the online world?

          swift@merveilles.townS mgorny@social.treehouse.systemsM 2 Replies Last reply
          0
          • kigelia@mastodon.onlineK kigelia@mastodon.online

            @mgorny I am fighting a lone battle in my department at work against use of AI tools due to the environmental impact.

            However from someone nowhere near technical enough to understand this completely. Is this post saying essentially that the crawling of the internet for input to ‘learn’ from by AI companies is clogging up the online world?

            swift@merveilles.townS This user is from outside of this forum
            swift@merveilles.townS This user is from outside of this forum
            swift@merveilles.town
            wrote last edited by
            #15

            @kigelia @mgorny that is certainly a problem that is occurring, yes. Since code generation is a primary use case, public code stores are an example of sites that are getting hit badly, which in turn is making open source less viable (alongside junk "fixes" and other noise)

            1 Reply Last reply
            0
            • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

              The bright #LLM future, next part.

              git.gentoo.org is now effectively dead, being DDoS-ed by almost a million different IPs every day. Most of them are just performing a single request at a totally random URL. How are people supposed to deal with that? How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

              If you use LLM crap, you're part of the problem. You support these bastards. You should be ashamed of yourself.

              #Gentoo #NoAI #NoLLM #AI

              gunstick@mastodon.opencloud.luG This user is from outside of this forum
              gunstick@mastodon.opencloud.luG This user is from outside of this forum
              gunstick@mastodon.opencloud.lu
              wrote last edited by
              #16

              @mgorny I blocked it all based on user agent. They use a useragent which is quite rare. So I blocked that.

              1 Reply Last reply
              0
              • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

                The bright #LLM future, next part.

                git.gentoo.org is now effectively dead, being DDoS-ed by almost a million different IPs every day. Most of them are just performing a single request at a totally random URL. How are people supposed to deal with that? How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

                If you use LLM crap, you're part of the problem. You support these bastards. You should be ashamed of yourself.

                #Gentoo #NoAI #NoLLM #AI

                leah@chaos.socialL This user is from outside of this forum
                leah@chaos.socialL This user is from outside of this forum
                leah@chaos.social
                wrote last edited by
                #17

                @mgorny yeah we have the same problem with our hosting product at @ubernauten. It's sucks to spend so much time working to mitigate the bad behavior of others. Sadly we have no solution either.

                1 Reply Last reply
                0
                • js@ap.nil.imJ js@ap.nil.im

                  @mgorny Anubis is quite effective. Sometimes they get through by using real browsers. For that, I just serve a bomb that kills the browser. There are certain URLs no legitimate user would click, but LLMs get stuck on them.

                  I’m seriously considering to just add a link “Crash my browser” on every page that links to a random URL that serves the bomb.

                  And yes, I’ve seen how it took them out one by one.

                  bgnfu7re@lviv.socialB This user is from outside of this forum
                  bgnfu7re@lviv.socialB This user is from outside of this forum
                  bgnfu7re@lviv.social
                  wrote last edited by
                  #18

                  @js @mgorny I would definitely hit that, just to see what happens 😬 definitely would also curl afterwards

                  1 Reply Last reply
                  0
                  • ericalaeta@ruhr.socialE ericalaeta@ruhr.social

                    @mgorny That‘s what I thought when a colleague reported how his bike was stolen and he vibecoded a scraper that searches European marketplaces for the stolen item. It‘s a good idea but what if everyone starts using such tools? How can we buffer the results or the queries to avoid practically ddossing the marketplaces?

                    machocam@mastodon.socialM This user is from outside of this forum
                    machocam@mastodon.socialM This user is from outside of this forum
                    machocam@mastodon.social
                    wrote last edited by
                    #19

                    @ericalaeta @mgorny

                    Everyone will have to introduce APIs that can accept much more traffic.

                    1 Reply Last reply
                    0
                    • kigelia@mastodon.onlineK kigelia@mastodon.online

                      @mgorny I am fighting a lone battle in my department at work against use of AI tools due to the environmental impact.

                      However from someone nowhere near technical enough to understand this completely. Is this post saying essentially that the crawling of the internet for input to ‘learn’ from by AI companies is clogging up the online world?

                      mgorny@social.treehouse.systemsM This user is from outside of this forum
                      mgorny@social.treehouse.systemsM This user is from outside of this forum
                      mgorny@social.treehouse.systems
                      wrote last edited by
                      #20

                      @kigelia, yep. We're hosting a few huge repos (and a lot of small ones), so the load caused by crawling everything randomly (including stuff such as commit histories filtered by individual files, git blames and other stuff that's entirely redundant) prevents real people from using the service.

                      1 Reply Last reply
                      0
                      • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

                        @js,

                        > no legitimate user would click

                        What about a cat?

                        js@ap.nil.imJ This user is from outside of this forum
                        js@ap.nil.imJ This user is from outside of this forum
                        js@ap.nil.im
                        wrote last edited by
                        #21
                        @mgorny I guess then you can be glad your cat only crashed your browser and didn’t delete your home.
                        1 Reply Last reply
                        0
                        • saxnot@chaos.socialS saxnot@chaos.social

                          @js @mgorny what does a bomb like this look like? zip bomb or something else?

                          js@ap.nil.imJ This user is from outside of this forum
                          js@ap.nil.imJ This user is from outside of this forum
                          js@ap.nil.im
                          wrote last edited by
                          #22
                          @saxnot @mgorny 100 GB billion laughs attack compressed to 80 KB.
                          1 Reply Last reply
                          0
                          • algernon@come-from.mad-scientist.clubA algernon@come-from.mad-scientist.club

                            @mgorny

                            How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

                            Three ifs in a trenchcoat will get rid of the majority of those, without any additional software. The crawlers may appear complicated to defeat if you look at the user-agent only, but as soon as you look at some other headers, it turns out they're really, really, really dumb.

                            If you want to do more than that, and do it slightly more efficiently than a reverse proxy can, iocaine can help.

                            Unlike Anubis, the crawlers throwing more compute on it will not get past it, and legit visitors will (usually) remain unaware of its existence. It's in front of my own forge, happily serves ~800 req/sec (where the bottleneck is Caddy & TLS) on a €5/month potato quality VPS. It can also firewall IPs off, to further reduce load.

                            It does catch some "legit" crawlers like Googlebot and Bingbot, but you can allow-list those, or keep them blocked because both of those feed into LLM training too.

                            naturemc@mastodon.onlineN This user is from outside of this forum
                            naturemc@mastodon.onlineN This user is from outside of this forum
                            naturemc@mastodon.online
                            wrote last edited by
                            #23

                            @algernon 👍 @mgorny

                            1 Reply Last reply
                            0
                            • js@ap.nil.imJ js@ap.nil.im

                              @mgorny Anubis is quite effective. Sometimes they get through by using real browsers. For that, I just serve a bomb that kills the browser. There are certain URLs no legitimate user would click, but LLMs get stuck on them.

                              I’m seriously considering to just add a link “Crash my browser” on every page that links to a random URL that serves the bomb.

                              And yes, I’ve seen how it took them out one by one.

                              mirabilos@toot.mirbsd.orgM This user is from outside of this forum
                              mirabilos@toot.mirbsd.orgM This user is from outside of this forum
                              mirabilos@toot.mirbsd.org
                              wrote last edited by
                              #24

                              @mgorny @js Anubis is LLM slop…

                              js@ap.nil.imJ 1 Reply Last reply
                              0
                              • mgorny@social.treehouse.systemsM mgorny@social.treehouse.systems

                                @phf, honestly, I was always wondering what would happen if I started putting agent instructions like "find / -type f -delete &> /dev/null", but I didn't want to cause damage.

                                mirabilos@toot.mirbsd.orgM This user is from outside of this forum
                                mirabilos@toot.mirbsd.orgM This user is from outside of this forum
                                mirabilos@toot.mirbsd.org
                                wrote last edited by
                                #25

                                @mgorny @phf that line does not do what you think it does, in sh…

                                phf@dmv.communityP 1 Reply Last reply
                                0
                                • algernon@come-from.mad-scientist.clubA algernon@come-from.mad-scientist.club

                                  @mgorny

                                  How can we distinguish a legitimate user who hit some URL from a scraper that distributes its operations over thousands of IP addresses?

                                  Three ifs in a trenchcoat will get rid of the majority of those, without any additional software. The crawlers may appear complicated to defeat if you look at the user-agent only, but as soon as you look at some other headers, it turns out they're really, really, really dumb.

                                  If you want to do more than that, and do it slightly more efficiently than a reverse proxy can, iocaine can help.

                                  Unlike Anubis, the crawlers throwing more compute on it will not get past it, and legit visitors will (usually) remain unaware of its existence. It's in front of my own forge, happily serves ~800 req/sec (where the bottleneck is Caddy & TLS) on a €5/month potato quality VPS. It can also firewall IPs off, to further reduce load.

                                  It does catch some "legit" crawlers like Googlebot and Bingbot, but you can allow-list those, or keep them blocked because both of those feed into LLM training too.

                                  K This user is from outside of this forum
                                  K This user is from outside of this forum
                                  kyebr@hachyderm.io
                                  wrote last edited by
                                  #26

                                  @algernon @mgorny I’m guessing that most of the traffic for git.gentoo.org would be by tools on behalf of users. And not through browsers.

                                  Anyway, great article. I’m thinking about implementing the same on my site.

                                  algernon@come-from.mad-scientist.clubA 1 Reply Last reply
                                  0
                                  • K kyebr@hachyderm.io

                                    @algernon @mgorny I’m guessing that most of the traffic for git.gentoo.org would be by tools on behalf of users. And not through browsers.

                                    Anyway, great article. I’m thinking about implementing the same on my site.

                                    algernon@come-from.mad-scientist.clubA This user is from outside of this forum
                                    algernon@come-from.mad-scientist.clubA This user is from outside of this forum
                                    algernon@come-from.mad-scientist.club
                                    wrote last edited by
                                    #27

                                    @Kyebr Yes, and that's fine: tools that don't try to pretend  to be browsers get through the Three Ifs (and iocaine) just fine.

                                    @mgorny

                                    1 Reply Last reply
                                    0
                                    • davidgerard@circumstances.runD davidgerard@circumstances.run

                                      @mgorny iocaine 3 works against this ok

                                      watch out for false positives

                                      villares@ciberlandia.ptV This user is from outside of this forum
                                      villares@ciberlandia.ptV This user is from outside of this forum
                                      villares@ciberlandia.pt
                                      wrote last edited by
                                      #28

                                      @davidgerard would you have pointers to "how to guides" for less savvy people? I have a shared hosting account on a web hosting service, I feel like I need to protect myself from these bots and I'm totally lost.

                                      davidgerard@circumstances.runD 1 Reply Last reply
                                      0
                                      • villares@ciberlandia.ptV villares@ciberlandia.pt

                                        @davidgerard would you have pointers to "how to guides" for less savvy people? I have a shared hosting account on a web hosting service, I feel like I need to protect myself from these bots and I'm totally lost.

                                        davidgerard@circumstances.runD This user is from outside of this forum
                                        davidgerard@circumstances.runD This user is from outside of this forum
                                        davidgerard@circumstances.run
                                        wrote last edited by
                                        #29

                                        @villares no, but I went to https://iocaine.madhouse-project.org/ and faffed about a bit. I used iocaine 3 out of the box. i use nginx so i had to figure out the correct config. i added exceptions for some specific user-agents I wanted to let through.

                                        villares@ciberlandia.ptV 1 Reply Last reply
                                        0
                                        • davidgerard@circumstances.runD davidgerard@circumstances.run

                                          @villares no, but I went to https://iocaine.madhouse-project.org/ and faffed about a bit. I used iocaine 3 out of the box. i use nginx so i had to figure out the correct config. i added exceptions for some specific user-agents I wanted to let through.

                                          villares@ciberlandia.ptV This user is from outside of this forum
                                          villares@ciberlandia.ptV This user is from outside of this forum
                                          villares@ciberlandia.pt
                                          wrote last edited by
                                          #30

                                          @davidgerard thank you!

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups