Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

Scheduled Pinned Locked Moved Uncategorized
39 Posts 26 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

    For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

    Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

    They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

    If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

    yora@mastodon.gamedev.placeY This user is from outside of this forum
    yora@mastodon.gamedev.placeY This user is from outside of this forum
    yora@mastodon.gamedev.place
    wrote last edited by
    #26

    @thomasfuchs Would Microsoft, Google, Facebook, and Nvidia lie to you?

    Yes, they do!

    1 Reply Last reply
    0
    • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

      For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

      Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

      They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

      If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

      zer0unplanned@friendica.rogueproject.orgZ This user is from outside of this forum
      zer0unplanned@friendica.rogueproject.orgZ This user is from outside of this forum
      zer0unplanned@friendica.rogueproject.org
      wrote last edited by
      #27
      @thomasfuchs I did not understood well, can you repeat for the 1001'th time please?
      1 Reply Last reply
      0
      • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

        For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

        Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

        They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

        If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

        L This user is from outside of this forum
        L This user is from outside of this forum
        libreovergratis@mastodon.social
        wrote last edited by
        #28

        @thomasfuchs Both sides of the AI debate are getting so insufferrable.

        If I see one more post about "It's just fancy autocomplete bro" I'm gonna freak.

        1 Reply Last reply
        0
        • cora@hachyderm.ioC cora@hachyderm.io

          @thomasfuchs Frankly I think it’s more plausible to describe the thought process of many humans in terms of token assemblage than the other way around.

          paxil@mastodon.socialP This user is from outside of this forum
          paxil@mastodon.socialP This user is from outside of this forum
          paxil@mastodon.social
          wrote last edited by
          #29

          @cora @thomasfuchs I would say parrot, AI, many humans in terms of assemblage, but it's close.

          1 Reply Last reply
          0
          • sinvega@mas.toS sinvega@mas.to

            @thomasfuchs I really, really wish people would stop with "hallucinated" when "fabricated" is both right there and more accurate

            tonyangelo@mspsocial.netT This user is from outside of this forum
            tonyangelo@mspsocial.netT This user is from outside of this forum
            tonyangelo@mspsocial.net
            wrote last edited by
            #30

            @sinvega @thomasfuchs “bullshit”
            https://en.wikipedia.org/wiki/On_Bullshit

            1 Reply Last reply
            0
            • madengineering@mastodon.cloudM madengineering@mastodon.cloud

              @thomasfuchs Lately they've taken the distinctly stupid idea of letting the chat bot effectively type commands directly into your shell and have them execute as if you typed them yourself, and just telling it not to type certain commands. Which it doesn't understand and does anyway.

              blotosmetek@circumstances.runB This user is from outside of this forum
              blotosmetek@circumstances.runB This user is from outside of this forum
              blotosmetek@circumstances.run
              wrote last edited by
              #31

              @madengineering @thomasfuchs …falling very much into the "destroy things" bin. So, yes, they can do that…

              1 Reply Last reply
              0
              • thomasfuchs@hachyderm.ioT This user is from outside of this forum
                thomasfuchs@hachyderm.ioT This user is from outside of this forum
                thomasfuchs@hachyderm.io
                wrote last edited by
                #32

                @hanscees I think I’ve seen some outside sometime

                1 Reply Last reply
                0
                • tambourineman@mastodon.cloudT tambourineman@mastodon.cloud

                  @thomasfuchs We don't know what makes one wake up in the morning and decide to climb a mountain or quit their job.
                  It may be some completely different process or there might be something to this pattern-matching statistical thing.
                  Do ants have agency? Do ant colonies?

                  We definitively must regulate the shit out of these big techs.
                  But saying that X does not do Y when both are poorly understood and defined is not the way, IMO.

                  thomasfuchs@hachyderm.ioT This user is from outside of this forum
                  thomasfuchs@hachyderm.ioT This user is from outside of this forum
                  thomasfuchs@hachyderm.io
                  wrote last edited by
                  #33

                  @tambourineman We know exactly how LLMs work, at every stage, literally humans created them.

                  They don’t have consciousness, they don’t have agency. They’re not even physical systems, so there is no self to realize.

                  Just because we don’t understand brains doesn’t mean we don’t understand some algorithm and hardware implementation for it.

                  tambourineman@mastodon.cloudT 1 Reply Last reply
                  0
                  • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                    @tambourineman We know exactly how LLMs work, at every stage, literally humans created them.

                    They don’t have consciousness, they don’t have agency. They’re not even physical systems, so there is no self to realize.

                    Just because we don’t understand brains doesn’t mean we don’t understand some algorithm and hardware implementation for it.

                    tambourineman@mastodon.cloudT This user is from outside of this forum
                    tambourineman@mastodon.cloudT This user is from outside of this forum
                    tambourineman@mastodon.cloud
                    wrote last edited by
                    #34

                    @thomasfuchs

                    Just because you build something doesn't mean you fully understand its implications. Emergent behavior exist, especially at this scale.
                    My point is that we don't need to get philosophical to criticize big tech.
                    They are destroying democracies, using our natural resources in a ponzi scheme that benefits very few at the detriment of billions, etc.
                    We have plenty of reasons for regulation already.

                    1 Reply Last reply
                    0
                    • tambourineman@mastodon.cloudT tambourineman@mastodon.cloud

                      @thomasfuchs We don't know what makes one wake up in the morning and decide to climb a mountain or quit their job.
                      It may be some completely different process or there might be something to this pattern-matching statistical thing.
                      Do ants have agency? Do ant colonies?

                      We definitively must regulate the shit out of these big techs.
                      But saying that X does not do Y when both are poorly understood and defined is not the way, IMO.

                      owlonabicycle@mastodon.worldO This user is from outside of this forum
                      owlonabicycle@mastodon.worldO This user is from outside of this forum
                      owlonabicycle@mastodon.world
                      wrote last edited by
                      #35

                      @tambourineman We obviously know that “X does not do Y” when it’s a machine, and we know exactly how it was programmed, and we know exactly what it’s doing. Everything about it is understood.

                      1 Reply Last reply
                      0
                      • williambob@mastodon.socialW This user is from outside of this forum
                        williambob@mastodon.socialW This user is from outside of this forum
                        williambob@mastodon.social
                        wrote last edited by
                        #36

                        @ClintonAnderson @thomasfuchs -nice to hear others knowing that making a machine in the image of of our minds. And FOMO is just fear which is the mindkiller

                        1 Reply Last reply
                        0
                        • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                          For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                          Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                          They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                          If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                          williambob@mastodon.socialW This user is from outside of this forum
                          williambob@mastodon.socialW This user is from outside of this forum
                          williambob@mastodon.social
                          wrote last edited by
                          #37

                          @thomasfuchs - old saying I half forget...as a computer in incapable of taking responsibility in any way, computer should never make management descisions

                          1 Reply Last reply
                          0
                          • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                            For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                            Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                            They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                            If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                            J This user is from outside of this forum
                            J This user is from outside of this forum
                            jeff_zucker@mastodon.social
                            wrote last edited by
                            #38

                            @thomasfuchs

                            Words matter. The goal of making us think of AI as a human being is woven into every interaction. For example :

                            Link Preview Image
                            1 Reply Last reply
                            0
                            • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                              For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                              Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                              They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                              If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                              crankylinuxuser@infosec.exchangeC This user is from outside of this forum
                              crankylinuxuser@infosec.exchangeC This user is from outside of this forum
                              crankylinuxuser@infosec.exchange
                              wrote last edited by
                              #39

                              @thomasfuchs

                              LLMs definitely can act. They can query the internet. They can use tools I teach them (MCP).

                              Do they think? I'm not particular sure that many humans even think. Or better yet, many humans respond in rote to the same stimuli (aka parse tokens and respond programmatically).

                              Given the recent neuroanatomy of LLMs, their findings are showing how LLMs start to work. What's surprising is that the starting circuits are decoding language, and the exiting circuits reencode language. And there appears to be a universal grammar (thanks Chomsky) internally, shared by many LLM models.

                              Link Preview Image
                              LLM Neuroanatomy: How I Topped the LLM Leaderboard Without Changing a Single Weight

                              ML, Biotech, Hardware, and Coordination Problems. Sometimes I write about hard problems and how to solve them.

                              favicon

                              David Noel Ng (dnhkng.github.io)

                              1 Reply Last reply
                              1
                              0
                              • R relay@relay.infosec.exchange shared this topic
                              Reply
                              • Reply as topic
                              Log in to reply
                              • Oldest to Newest
                              • Newest to Oldest
                              • Most Votes


                              • Login

                              • Login or register to search.
                              • First post
                                Last post
                              0
                              • Categories
                              • Recent
                              • Tags
                              • Popular
                              • World
                              • Users
                              • Groups