Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

Scheduled Pinned Locked Moved Uncategorized
39 Posts 26 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

    For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

    Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

    They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

    If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

    madengineering@mastodon.cloudM This user is from outside of this forum
    madengineering@mastodon.cloudM This user is from outside of this forum
    madengineering@mastodon.cloud
    wrote last edited by
    #2

    @thomasfuchs Lately they've taken the distinctly stupid idea of letting the chat bot effectively type commands directly into your shell and have them execute as if you typed them yourself, and just telling it not to type certain commands. Which it doesn't understand and does anyway.

    blotosmetek@circumstances.runB 1 Reply Last reply
    0
    • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

      For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

      Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

      They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

      If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

      sinvega@mas.toS This user is from outside of this forum
      sinvega@mas.toS This user is from outside of this forum
      sinvega@mas.to
      wrote last edited by
      #3

      @thomasfuchs I really, really wish people would stop with "hallucinated" when "fabricated" is both right there and more accurate

      nimro@hachyderm.ioN tonyangelo@mspsocial.netT 2 Replies Last reply
      0
      • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

        For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

        Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

        They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

        If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

        cora@hachyderm.ioC This user is from outside of this forum
        cora@hachyderm.ioC This user is from outside of this forum
        cora@hachyderm.io
        wrote last edited by
        #4

        @thomasfuchs Frankly I think it’s more plausible to describe the thought process of many humans in terms of token assemblage than the other way around.

        paxil@mastodon.socialP 1 Reply Last reply
        0
        • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

          For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

          Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

          They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

          If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

          michaelgemar@cosocial.caM This user is from outside of this forum
          michaelgemar@cosocial.caM This user is from outside of this forum
          michaelgemar@cosocial.ca
          wrote last edited by
          #5

          @thomasfuchs @WeirdWriter I really think that regulations should insist that LLMs software be configured to not refer to “itself” with personal pronouns, or imply it has emotional states, or all the other rhetorical tricks they have been programmed to use to appear “human”.

          thomasfuchs@hachyderm.ioT 1 Reply Last reply
          0
          • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

            For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

            Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

            They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

            If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

            tambourineman@mastodon.cloudT This user is from outside of this forum
            tambourineman@mastodon.cloudT This user is from outside of this forum
            tambourineman@mastodon.cloud
            wrote last edited by
            #6

            @thomasfuchs We don't know what makes one wake up in the morning and decide to climb a mountain or quit their job.
            It may be some completely different process or there might be something to this pattern-matching statistical thing.
            Do ants have agency? Do ant colonies?

            We definitively must regulate the shit out of these big techs.
            But saying that X does not do Y when both are poorly understood and defined is not the way, IMO.

            thomasfuchs@hachyderm.ioT owlonabicycle@mastodon.worldO 2 Replies Last reply
            0
            • sinvega@mas.toS sinvega@mas.to

              @thomasfuchs I really, really wish people would stop with "hallucinated" when "fabricated" is both right there and more accurate

              nimro@hachyderm.ioN This user is from outside of this forum
              nimro@hachyderm.ioN This user is from outside of this forum
              nimro@hachyderm.io
              wrote last edited by
              #7

              @sinvega this paper makes a compelling case for using the academic term “bullshit” https://arxiv.org/abs/2507.07484

              1 Reply Last reply
              0
              • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                frog_reborn@mstdn.socialF This user is from outside of this forum
                frog_reborn@mstdn.socialF This user is from outside of this forum
                frog_reborn@mstdn.social
                wrote last edited by
                #8

                @thomasfuchs

                The first two don't really make sense to me. A virus can "evade safeguards" and a meteorite can "destroy things", so I don't think there has to be much agency involved in the first place.

                The latter seems more like a more fitting criticism, but in all three cases I'm also not sure how one were to phrase it alternatively.

                thomasfuchs@hachyderm.ioT 1 Reply Last reply
                0
                • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                  For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                  Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                  They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                  If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                  eric_neue@indieweb.socialE This user is from outside of this forum
                  eric_neue@indieweb.socialE This user is from outside of this forum
                  eric_neue@indieweb.social
                  wrote last edited by
                  #9

                  @thomasfuchs i wish we could educate the public that LLMs would be more accurately described as “simulated intelligence,” but i can’t figure out how to explain the difference to normies at all.

                  1 Reply Last reply
                  0
                  • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                    For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                    Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                    They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                    If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                    S This user is from outside of this forum
                    S This user is from outside of this forum
                    slotos@toot.community
                    wrote last edited by
                    #10

                    @thomasfuchs You don’t need agency to evade safeguards, destroy things, or ignore instructions. `rm` can do it.

                    This is literally the mistake people you criticize are making - imbuing intent where there’s none.

                    The underlying tech had been apt at finding ways to circumvent feedback loops since before the bubble. This is constrained to the training phase, but with verification of commercial models being mathematically infeasible, these avoidance patterns are shipped directly to users.

                    thomasfuchs@hachyderm.ioT 1 Reply Last reply
                    0
                    • R relay@relay.publicsquare.global shared this topic
                    • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                      For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                      Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                      They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                      If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                      brian@social.brian.jpB This user is from outside of this forum
                      brian@social.brian.jpB This user is from outside of this forum
                      brian@social.brian.jp
                      wrote last edited by
                      #11

                      @shimst3r You’re absolutely right!

                      1 Reply Last reply
                      0
                      • S slotos@toot.community

                        @thomasfuchs You don’t need agency to evade safeguards, destroy things, or ignore instructions. `rm` can do it.

                        This is literally the mistake people you criticize are making - imbuing intent where there’s none.

                        The underlying tech had been apt at finding ways to circumvent feedback loops since before the bubble. This is constrained to the training phase, but with verification of commercial models being mathematically infeasible, these avoidance patterns are shipped directly to users.

                        thomasfuchs@hachyderm.ioT This user is from outside of this forum
                        thomasfuchs@hachyderm.ioT This user is from outside of this forum
                        thomasfuchs@hachyderm.io
                        wrote last edited by
                        #12

                        @slotos My point is that using active verbs like “evade” is misleading (yourself and others), it implies purpose in choosing and pursuing an action.

                        LLMs do not actively chose to do anything.

                        S 1 Reply Last reply
                        0
                        • michaelgemar@cosocial.caM michaelgemar@cosocial.ca

                          @thomasfuchs @WeirdWriter I really think that regulations should insist that LLMs software be configured to not refer to “itself” with personal pronouns, or imply it has emotional states, or all the other rhetorical tricks they have been programmed to use to appear “human”.

                          thomasfuchs@hachyderm.ioT This user is from outside of this forum
                          thomasfuchs@hachyderm.ioT This user is from outside of this forum
                          thomasfuchs@hachyderm.io
                          wrote last edited by
                          #13

                          @michaelgemar @WeirdWriter Yes anthropomorphized chatbots should be illegal.

                          There’s plenty of other ways to interact with LLMs that don’t cause psychosis (for example autocomplete of whole sentences, something that can be useful for things like coding.)

                          elricofmelnibone@mastodon.socialE 1 Reply Last reply
                          0
                          • frog_reborn@mstdn.socialF frog_reborn@mstdn.social

                            @thomasfuchs

                            The first two don't really make sense to me. A virus can "evade safeguards" and a meteorite can "destroy things", so I don't think there has to be much agency involved in the first place.

                            The latter seems more like a more fitting criticism, but in all three cases I'm also not sure how one were to phrase it alternatively.

                            thomasfuchs@hachyderm.ioT This user is from outside of this forum
                            thomasfuchs@hachyderm.ioT This user is from outside of this forum
                            thomasfuchs@hachyderm.io
                            wrote last edited by
                            #14

                            @frog_reborn a virus has evolved to evade—it’s actively doing evasion, purposefully.

                            Destroy has multiple meanings as a verb, but when used with what LLMs do people mean on purpose; as opposed to accidentally damaging something.

                            frog_reborn@mstdn.socialF 1 Reply Last reply
                            0
                            • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                              For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                              Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                              They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                              If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                              wolf4earth@hachyderm.ioW This user is from outside of this forum
                              wolf4earth@hachyderm.ioW This user is from outside of this forum
                              wolf4earth@hachyderm.io
                              wrote last edited by
                              #15

                              @thomasfuchs I don't disagree. AI is a statistical mirror. And I believe your take is reductionist. Let me be a bit provocative:

                              For the 1,000th time: "Humans" don't have agency and cannot actually decide anything.

                              They literally only do one thing and one thing only: reproduce neurochemical chain reactions based on pre-existing connectivity between synapses in a nervous system.

                              If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely touch grass.

                              ---

                              Do I believe AI has agency? No, not yet.
                              Do I believe people have agency? Yes.
                              Do I believe people severely underestimate how much we reproduce neurological conditioning? Yes.

                              Both produce statistical inference. Only one can currently modify their own constraints.

                              Not equivalent. Not nothing.

                              clintruin@mastodon.socialC 1 Reply Last reply
                              0
                              • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                                @slotos My point is that using active verbs like “evade” is misleading (yourself and others), it implies purpose in choosing and pursuing an action.

                                LLMs do not actively chose to do anything.

                                S This user is from outside of this forum
                                S This user is from outside of this forum
                                slotos@toot.community
                                wrote last edited by
                                #16

                                @thomasfuchs That’s a general natural language problem.

                                For example, „you’re avoiding responsibility” and „he avoided responsibility” use the same verb with very different connotations when it comes to intent attribution.

                                Our verbs aren’t that clear cut on their own. We also tend to merge or specialize closely related ones.

                                That is a reason why `AGENTS.md` is a braindead idea, for example. But that’s a separate rant entirely.

                                thomasfuchs@hachyderm.ioT 1 Reply Last reply
                                0
                                • S slotos@toot.community

                                  @thomasfuchs That’s a general natural language problem.

                                  For example, „you’re avoiding responsibility” and „he avoided responsibility” use the same verb with very different connotations when it comes to intent attribution.

                                  Our verbs aren’t that clear cut on their own. We also tend to merge or specialize closely related ones.

                                  That is a reason why `AGENTS.md` is a braindead idea, for example. But that’s a separate rant entirely.

                                  thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                  thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                  thomasfuchs@hachyderm.io
                                  wrote last edited by
                                  #17

                                  @slotos Perhaps, but using literally any verb with what LLMs generate other than “generate” is misleading.

                                  You wouldn’t call your dice “evading” if you use them to randomly select some nouns and verbs from a dictionary and it happens to say “lie about deleting the root folder”.

                                  S 1 Reply Last reply
                                  0
                                  • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                                    @michaelgemar @WeirdWriter Yes anthropomorphized chatbots should be illegal.

                                    There’s plenty of other ways to interact with LLMs that don’t cause psychosis (for example autocomplete of whole sentences, something that can be useful for things like coding.)

                                    elricofmelnibone@mastodon.socialE This user is from outside of this forum
                                    elricofmelnibone@mastodon.socialE This user is from outside of this forum
                                    elricofmelnibone@mastodon.social
                                    wrote last edited by
                                    #18

                                    @thomasfuchs Autocompleting whole sentences is just as bad. How do you know that sentence is what you wanted to write in the first place?

                                    thomasfuchs@hachyderm.ioT 1 Reply Last reply
                                    0
                                    • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                                      For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                                      Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                                      They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                                      If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                                      keydelk@fosstodon.orgK This user is from outside of this forum
                                      keydelk@fosstodon.orgK This user is from outside of this forum
                                      keydelk@fosstodon.org
                                      wrote last edited by
                                      #19

                                      @thomasfuchs tech bros be like “but what if we call it ‘agentic AI’ and pipe the output of the plausible sentence generator straight into the bash shell (and give it sudo privileges for good measure)”

                                      1 Reply Last reply
                                      0
                                      • elricofmelnibone@mastodon.socialE elricofmelnibone@mastodon.social

                                        @thomasfuchs Autocompleting whole sentences is just as bad. How do you know that sentence is what you wanted to write in the first place?

                                        thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                        thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                        thomasfuchs@hachyderm.io
                                        wrote last edited by
                                        #20

                                        @elricofmelnibone you see it while your typing so you know if it’s what you wanted?

                                        this can be helpful especially for people who can’t type fast and to avoid common typos ¯\_(ツ)_/¯

                                        it’s nothing like “just as bad” as a sycophantic chatbot that constantly brownnoses you

                                        1 Reply Last reply
                                        0
                                        • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                                          @frog_reborn a virus has evolved to evade—it’s actively doing evasion, purposefully.

                                          Destroy has multiple meanings as a verb, but when used with what LLMs do people mean on purpose; as opposed to accidentally damaging something.

                                          frog_reborn@mstdn.socialF This user is from outside of this forum
                                          frog_reborn@mstdn.socialF This user is from outside of this forum
                                          frog_reborn@mstdn.social
                                          wrote last edited by
                                          #21

                                          @thomasfuchs

                                          "a virus has evolved to evade—it’s actively doing evasion, purposefully."

                                          That's an opinion that's pretty firmly outside the biological mainstream.

                                          (Our biology teacher would always scold us everytime one of said "X evolved to do Y")

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups