Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

Scheduled Pinned Locked Moved Uncategorized
39 Posts 26 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • S slotos@toot.community

    @thomasfuchs That’s a general natural language problem.

    For example, „you’re avoiding responsibility” and „he avoided responsibility” use the same verb with very different connotations when it comes to intent attribution.

    Our verbs aren’t that clear cut on their own. We also tend to merge or specialize closely related ones.

    That is a reason why `AGENTS.md` is a braindead idea, for example. But that’s a separate rant entirely.

    thomasfuchs@hachyderm.ioT This user is from outside of this forum
    thomasfuchs@hachyderm.ioT This user is from outside of this forum
    thomasfuchs@hachyderm.io
    wrote last edited by
    #17

    @slotos Perhaps, but using literally any verb with what LLMs generate other than “generate” is misleading.

    You wouldn’t call your dice “evading” if you use them to randomly select some nouns and verbs from a dictionary and it happens to say “lie about deleting the root folder”.

    S 1 Reply Last reply
    0
    • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

      @michaelgemar @WeirdWriter Yes anthropomorphized chatbots should be illegal.

      There’s plenty of other ways to interact with LLMs that don’t cause psychosis (for example autocomplete of whole sentences, something that can be useful for things like coding.)

      elricofmelnibone@mastodon.socialE This user is from outside of this forum
      elricofmelnibone@mastodon.socialE This user is from outside of this forum
      elricofmelnibone@mastodon.social
      wrote last edited by
      #18

      @thomasfuchs Autocompleting whole sentences is just as bad. How do you know that sentence is what you wanted to write in the first place?

      thomasfuchs@hachyderm.ioT 1 Reply Last reply
      0
      • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

        For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

        Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

        They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

        If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

        keydelk@fosstodon.orgK This user is from outside of this forum
        keydelk@fosstodon.orgK This user is from outside of this forum
        keydelk@fosstodon.org
        wrote last edited by
        #19

        @thomasfuchs tech bros be like “but what if we call it ‘agentic AI’ and pipe the output of the plausible sentence generator straight into the bash shell (and give it sudo privileges for good measure)”

        1 Reply Last reply
        0
        • elricofmelnibone@mastodon.socialE elricofmelnibone@mastodon.social

          @thomasfuchs Autocompleting whole sentences is just as bad. How do you know that sentence is what you wanted to write in the first place?

          thomasfuchs@hachyderm.ioT This user is from outside of this forum
          thomasfuchs@hachyderm.ioT This user is from outside of this forum
          thomasfuchs@hachyderm.io
          wrote last edited by
          #20

          @elricofmelnibone you see it while your typing so you know if it’s what you wanted?

          this can be helpful especially for people who can’t type fast and to avoid common typos ¯\_(ツ)_/¯

          it’s nothing like “just as bad” as a sycophantic chatbot that constantly brownnoses you

          1 Reply Last reply
          0
          • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

            @frog_reborn a virus has evolved to evade—it’s actively doing evasion, purposefully.

            Destroy has multiple meanings as a verb, but when used with what LLMs do people mean on purpose; as opposed to accidentally damaging something.

            frog_reborn@mstdn.socialF This user is from outside of this forum
            frog_reborn@mstdn.socialF This user is from outside of this forum
            frog_reborn@mstdn.social
            wrote last edited by
            #21

            @thomasfuchs

            "a virus has evolved to evade—it’s actively doing evasion, purposefully."

            That's an opinion that's pretty firmly outside the biological mainstream.

            (Our biology teacher would always scold us everytime one of said "X evolved to do Y")

            1 Reply Last reply
            0
            • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

              @slotos Perhaps, but using literally any verb with what LLMs generate other than “generate” is misleading.

              You wouldn’t call your dice “evading” if you use them to randomly select some nouns and verbs from a dictionary and it happens to say “lie about deleting the root folder”.

              S This user is from outside of this forum
              S This user is from outside of this forum
              slotos@toot.community
              wrote last edited by
              #22

              @thomasfuchs It’s has been a useful way to describe things. We use those same verbs to describe behavior of malware without any issues.

              The problem arises not from the verbs themselves, but from the targeted campaign to establish a false premise that AI has agency [and will doom us all].

              It’s not that these verbs imply agency, but that the pool is so poisoned that the usual verbs fail due to implied agency.

              Which is a long way to say „I concede your point”.

              thomasfuchs@hachyderm.ioT 1 Reply Last reply
              0
              • wolf4earth@hachyderm.ioW wolf4earth@hachyderm.io

                @thomasfuchs I don't disagree. AI is a statistical mirror. And I believe your take is reductionist. Let me be a bit provocative:

                For the 1,000th time: "Humans" don't have agency and cannot actually decide anything.

                They literally only do one thing and one thing only: reproduce neurochemical chain reactions based on pre-existing connectivity between synapses in a nervous system.

                If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely touch grass.

                ---

                Do I believe AI has agency? No, not yet.
                Do I believe people have agency? Yes.
                Do I believe people severely underestimate how much we reproduce neurological conditioning? Yes.

                Both produce statistical inference. Only one can currently modify their own constraints.

                Not equivalent. Not nothing.

                clintruin@mastodon.socialC This user is from outside of this forum
                clintruin@mastodon.socialC This user is from outside of this forum
                clintruin@mastodon.social
                wrote last edited by
                #23

                @wolf4earth @thomasfuchs
                "Nonexistence never hurt anyone. Existence hurts everyone."
                - Thomas Ligotti

                1 Reply Last reply
                0
                • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                  For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                  Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                  They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                  If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                  jsc@hcommons.socialJ This user is from outside of this forum
                  jsc@hcommons.socialJ This user is from outside of this forum
                  jsc@hcommons.social
                  wrote last edited by
                  #24

                  @thomasfuchs A thousand times "yes" to your ostensibly thousandth time uttering this truth. Anyone who's paying attention recognizes that computers are necessarily deterministic by design and words like "AI', "agency", and "hallucinate" are at best shorthand for observed operations, and at worst, deceptive marketing terms.

                  1 Reply Last reply
                  0
                  • S slotos@toot.community

                    @thomasfuchs It’s has been a useful way to describe things. We use those same verbs to describe behavior of malware without any issues.

                    The problem arises not from the verbs themselves, but from the targeted campaign to establish a false premise that AI has agency [and will doom us all].

                    It’s not that these verbs imply agency, but that the pool is so poisoned that the usual verbs fail due to implied agency.

                    Which is a long way to say „I concede your point”.

                    thomasfuchs@hachyderm.ioT This user is from outside of this forum
                    thomasfuchs@hachyderm.ioT This user is from outside of this forum
                    thomasfuchs@hachyderm.io
                    wrote last edited by
                    #25

                    @slotos I think I agree. Fwiw for malware it’s more like “the human who wrote it purposefully planned it such that it can evade e.g. a virus scanner”

                    This can be true for AI-generated code etc as well (steered there by prompts) but my OP was talking about sort of self-arising actions (which don’t exist).

                    1 Reply Last reply
                    0
                    • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                      For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                      Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                      They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                      If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                      yora@mastodon.gamedev.placeY This user is from outside of this forum
                      yora@mastodon.gamedev.placeY This user is from outside of this forum
                      yora@mastodon.gamedev.place
                      wrote last edited by
                      #26

                      @thomasfuchs Would Microsoft, Google, Facebook, and Nvidia lie to you?

                      Yes, they do!

                      1 Reply Last reply
                      0
                      • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                        For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                        Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                        They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                        If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                        zer0unplanned@friendica.rogueproject.orgZ This user is from outside of this forum
                        zer0unplanned@friendica.rogueproject.orgZ This user is from outside of this forum
                        zer0unplanned@friendica.rogueproject.org
                        wrote last edited by
                        #27
                        @thomasfuchs I did not understood well, can you repeat for the 1001'th time please?
                        1 Reply Last reply
                        0
                        • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                          For the 1,000th time: "AI" does not have agency and cannot think and cannot act.

                          Chatbots cannot "evade safeguards" or "destroy things" or "ignore instructions".

                          They do literally only do one thing and one thing only: string tokens together based on statistics of proximity of tokens in a data corpus.

                          If you attribute any deeper meaning to this, it's a sign of psychosis and you should absolutely never use chatbots, possibly you should even touch grass.

                          L This user is from outside of this forum
                          L This user is from outside of this forum
                          libreovergratis@mastodon.social
                          wrote last edited by
                          #28

                          @thomasfuchs Both sides of the AI debate are getting so insufferrable.

                          If I see one more post about "It's just fancy autocomplete bro" I'm gonna freak.

                          1 Reply Last reply
                          0
                          • cora@hachyderm.ioC cora@hachyderm.io

                            @thomasfuchs Frankly I think it’s more plausible to describe the thought process of many humans in terms of token assemblage than the other way around.

                            paxil@mastodon.socialP This user is from outside of this forum
                            paxil@mastodon.socialP This user is from outside of this forum
                            paxil@mastodon.social
                            wrote last edited by
                            #29

                            @cora @thomasfuchs I would say parrot, AI, many humans in terms of assemblage, but it's close.

                            1 Reply Last reply
                            0
                            • sinvega@mas.toS sinvega@mas.to

                              @thomasfuchs I really, really wish people would stop with "hallucinated" when "fabricated" is both right there and more accurate

                              tonyangelo@mspsocial.netT This user is from outside of this forum
                              tonyangelo@mspsocial.netT This user is from outside of this forum
                              tonyangelo@mspsocial.net
                              wrote last edited by
                              #30

                              @sinvega @thomasfuchs “bullshit”
                              https://en.wikipedia.org/wiki/On_Bullshit

                              1 Reply Last reply
                              0
                              • madengineering@mastodon.cloudM madengineering@mastodon.cloud

                                @thomasfuchs Lately they've taken the distinctly stupid idea of letting the chat bot effectively type commands directly into your shell and have them execute as if you typed them yourself, and just telling it not to type certain commands. Which it doesn't understand and does anyway.

                                blotosmetek@circumstances.runB This user is from outside of this forum
                                blotosmetek@circumstances.runB This user is from outside of this forum
                                blotosmetek@circumstances.run
                                wrote last edited by
                                #31

                                @madengineering @thomasfuchs …falling very much into the "destroy things" bin. So, yes, they can do that…

                                1 Reply Last reply
                                0
                                • thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                  thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                  thomasfuchs@hachyderm.io
                                  wrote last edited by
                                  #32

                                  @hanscees I think I’ve seen some outside sometime

                                  1 Reply Last reply
                                  0
                                  • tambourineman@mastodon.cloudT tambourineman@mastodon.cloud

                                    @thomasfuchs We don't know what makes one wake up in the morning and decide to climb a mountain or quit their job.
                                    It may be some completely different process or there might be something to this pattern-matching statistical thing.
                                    Do ants have agency? Do ant colonies?

                                    We definitively must regulate the shit out of these big techs.
                                    But saying that X does not do Y when both are poorly understood and defined is not the way, IMO.

                                    thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                    thomasfuchs@hachyderm.ioT This user is from outside of this forum
                                    thomasfuchs@hachyderm.io
                                    wrote last edited by
                                    #33

                                    @tambourineman We know exactly how LLMs work, at every stage, literally humans created them.

                                    They don’t have consciousness, they don’t have agency. They’re not even physical systems, so there is no self to realize.

                                    Just because we don’t understand brains doesn’t mean we don’t understand some algorithm and hardware implementation for it.

                                    tambourineman@mastodon.cloudT 1 Reply Last reply
                                    0
                                    • thomasfuchs@hachyderm.ioT thomasfuchs@hachyderm.io

                                      @tambourineman We know exactly how LLMs work, at every stage, literally humans created them.

                                      They don’t have consciousness, they don’t have agency. They’re not even physical systems, so there is no self to realize.

                                      Just because we don’t understand brains doesn’t mean we don’t understand some algorithm and hardware implementation for it.

                                      tambourineman@mastodon.cloudT This user is from outside of this forum
                                      tambourineman@mastodon.cloudT This user is from outside of this forum
                                      tambourineman@mastodon.cloud
                                      wrote last edited by
                                      #34

                                      @thomasfuchs

                                      Just because you build something doesn't mean you fully understand its implications. Emergent behavior exist, especially at this scale.
                                      My point is that we don't need to get philosophical to criticize big tech.
                                      They are destroying democracies, using our natural resources in a ponzi scheme that benefits very few at the detriment of billions, etc.
                                      We have plenty of reasons for regulation already.

                                      1 Reply Last reply
                                      0
                                      • tambourineman@mastodon.cloudT tambourineman@mastodon.cloud

                                        @thomasfuchs We don't know what makes one wake up in the morning and decide to climb a mountain or quit their job.
                                        It may be some completely different process or there might be something to this pattern-matching statistical thing.
                                        Do ants have agency? Do ant colonies?

                                        We definitively must regulate the shit out of these big techs.
                                        But saying that X does not do Y when both are poorly understood and defined is not the way, IMO.

                                        owlonabicycle@mastodon.worldO This user is from outside of this forum
                                        owlonabicycle@mastodon.worldO This user is from outside of this forum
                                        owlonabicycle@mastodon.world
                                        wrote last edited by
                                        #35

                                        @tambourineman We obviously know that “X does not do Y” when it’s a machine, and we know exactly how it was programmed, and we know exactly what it’s doing. Everything about it is understood.

                                        1 Reply Last reply
                                        0
                                        • williambob@mastodon.socialW This user is from outside of this forum
                                          williambob@mastodon.socialW This user is from outside of this forum
                                          williambob@mastodon.social
                                          wrote last edited by
                                          #36

                                          @ClintonAnderson @thomasfuchs -nice to hear others knowing that making a machine in the image of of our minds. And FOMO is just fear which is the mindkiller

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups