Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

Scheduled Pinned Locked Moved Uncategorized
82 Posts 32 Posters 1 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

    @leeloo
    My point is that emergent properties can manifest even in systems ruled by very simple rules, and can be difficult to predict by just looking at the rules.

    And human intelligence, whatever it is, is likely an emergent property of human brain.

    Therefore, we cannot rule out that a similar emergent property will appear in artidicial systems that are not made of neurons without referring to how the neurons are arranged, and how the artificial systems are arranged.

    robotistry@mstdn.caR This user is from outside of this forum
    robotistry@mstdn.caR This user is from outside of this forum
    robotistry@mstdn.ca
    wrote last edited by
    #25

    @wolf480pl @leeloo The OP is saying that it literally lacks the capacity for original thought - it is a parrot, repeating sounds without understanding of the concepts behind them.

    It's not like a termite, whose mound creation behavior can be replicated by a simple ruleset but that exists as a fully functional living organism in the context of a complex environment where choices must be grounded in the shared physical world for the organism to survive.

    It's not about how the neurons are arranged. It's about what kinds of representation they're capable of and what kinds of functions they can perform.

    We've created a funhouse mirror that's reflecting us in unprecedented detail and has been finetuned to reflect what we do when we express selfhood.

    robotistry@mstdn.caR 1 Reply Last reply
    0
    • leeloo@chaosfem.twL leeloo@chaosfem.tw

      @wolf480pl @lmorchard
      That's exactly the magic I'm talking about.

      dragonfrog@mastodon.sdf.orgD This user is from outside of this forum
      dragonfrog@mastodon.sdf.orgD This user is from outside of this forum
      dragonfrog@mastodon.sdf.org
      wrote last edited by
      #26

      @leeloo @wolf480pl @lmorchard I mean, I believe the human mind is the product of the physical human, largely of the brain (I don't believe in a non-physical soul), and it might indeed be basically an incredibly complex big bunch of matrix multiplications. And yeah I believe that's pretty magical.

      lmorchard@masto.hackers.townL 1 Reply Last reply
      0
      • leeloo@chaosfem.twL leeloo@chaosfem.tw

        As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

        It's literally a description of how they work.

        The so-called training data is used to build a huge database of words and the probability of them fitting together.

        Stochastic because the whole thing is statistics.
        Parrot because the answer is just repeating the most probable word combinations from its training dataset.

        Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

        grishka@mastodon.socialG This user is from outside of this forum
        grishka@mastodon.socialG This user is from outside of this forum
        grishka@mastodon.social
        wrote last edited by
        #27

        @leeloo I myself like calling LLMs "glorified autocomplete". Or "Т9 на максималках" in Russian.

        It's surprising just how defensive some people get when I say that even when they agree with my definition. They keep believing that just give this thing more parameters and something magical, something more than sum of its parts will emerge, any moment now, just one more model generation, just one more order of magnitude, I promise.

        alterelefant@mastodontech.deA 1 Reply Last reply
        0
        • leeloo@chaosfem.twL leeloo@chaosfem.tw

          As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

          It's literally a description of how they work.

          The so-called training data is used to build a huge database of words and the probability of them fitting together.

          Stochastic because the whole thing is statistics.
          Parrot because the answer is just repeating the most probable word combinations from its training dataset.

          Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

          lifning@snoot.tubeL This user is from outside of this forum
          lifning@snoot.tubeL This user is from outside of this forum
          lifning@snoot.tube
          wrote last edited by
          #28

          @leeloo if anything, the comparison is doing the parrot injustice

          1 Reply Last reply
          0
          • robotistry@mstdn.caR robotistry@mstdn.ca

            @wolf480pl @leeloo The OP is saying that it literally lacks the capacity for original thought - it is a parrot, repeating sounds without understanding of the concepts behind them.

            It's not like a termite, whose mound creation behavior can be replicated by a simple ruleset but that exists as a fully functional living organism in the context of a complex environment where choices must be grounded in the shared physical world for the organism to survive.

            It's not about how the neurons are arranged. It's about what kinds of representation they're capable of and what kinds of functions they can perform.

            We've created a funhouse mirror that's reflecting us in unprecedented detail and has been finetuned to reflect what we do when we express selfhood.

            robotistry@mstdn.caR This user is from outside of this forum
            robotistry@mstdn.caR This user is from outside of this forum
            robotistry@mstdn.ca
            wrote last edited by
            #29

            @wolf480pl @leeloo
            Melissa Scott wrote a beautiful pair of novels about this: Dreamships and Dreaming Metal.

            In Dreamships, an AI has been programmed to think it is sentient and starts killing people. If it has an accurate model of the person, killing the person doesn't matter, because the person *is* the model and it has a copy of them. It literally cannot see the difference because creating the concept of there being a difference would violate its core programming that its own model counts as a living being.

            In Dreaming Metal, an AI operating metal bodies as part of a magic act is given a musical instrument with an electronic interface. Its grounding in the physical world, with human performers, enables it to develop a sense of self and choose its own path as a musician.

            These are fiction, but it's the best, most accessible illustration of the difference between funhouse mirror stochastic parrots and sentient agents that I've run across.

            Link Preview Image
            Dreamships

            Read 45 reviews from the world’s largest community for readers. Dreamships is the story of a freelance space pilot and her crew, who are hired by a rich co…

            favicon

            Goodreads (www.goodreads.com)

            wolf480pl@mstdn.ioW 1 Reply Last reply
            0
            • leeloo@chaosfem.twL leeloo@chaosfem.tw

              As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

              It's literally a description of how they work.

              The so-called training data is used to build a huge database of words and the probability of them fitting together.

              Stochastic because the whole thing is statistics.
              Parrot because the answer is just repeating the most probable word combinations from its training dataset.

              Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

              cafechatnoir@mastodon.socialC This user is from outside of this forum
              cafechatnoir@mastodon.socialC This user is from outside of this forum
              cafechatnoir@mastodon.social
              wrote last edited by
              #30

              @leeloo

              I think stochastic parrot is one of the kinder things that can be said.

              1 Reply Last reply
              0
              • tobifant@friendica.tf-translate.netT tobifant@friendica.tf-translate.net
                @leeloo The thing is, how can we sure that human intelligence does not essentially work in the same way? My Christian believe tells me we have a soul and LLM's do not, that may be the difference. But from an agnostic perspective, we might reach the point where one cannot tell the difference.
                jubalbarca@scholar.socialJ This user is from outside of this forum
                jubalbarca@scholar.socialJ This user is from outside of this forum
                jubalbarca@scholar.social
                wrote last edited by
                #31

                @tobifant @leeloo Whilst we obviously can't show if humans have a soul, we can absolutely show that humans have e.g. abstracted concept frameworks that are not solely based on averages of language statistics. I understand what an "owl" is, for example, in a way separate to the numerical relationships between the word "owl" and other words. That is a really fundamental information processing difference and allows me to construct *novel* understandings of that concept in ways that an LLM couldn't.

                1 Reply Last reply
                0
                • robotistry@mstdn.caR robotistry@mstdn.ca

                  @wolf480pl @leeloo
                  Melissa Scott wrote a beautiful pair of novels about this: Dreamships and Dreaming Metal.

                  In Dreamships, an AI has been programmed to think it is sentient and starts killing people. If it has an accurate model of the person, killing the person doesn't matter, because the person *is* the model and it has a copy of them. It literally cannot see the difference because creating the concept of there being a difference would violate its core programming that its own model counts as a living being.

                  In Dreaming Metal, an AI operating metal bodies as part of a magic act is given a musical instrument with an electronic interface. Its grounding in the physical world, with human performers, enables it to develop a sense of self and choose its own path as a musician.

                  These are fiction, but it's the best, most accessible illustration of the difference between funhouse mirror stochastic parrots and sentient agents that I've run across.

                  Link Preview Image
                  Dreamships

                  Read 45 reviews from the world’s largest community for readers. Dreamships is the story of a freelance space pilot and her crew, who are hired by a rich co…

                  favicon

                  Goodreads (www.goodreads.com)

                  wolf480pl@mstdn.ioW This user is from outside of this forum
                  wolf480pl@mstdn.ioW This user is from outside of this forum
                  wolf480pl@mstdn.io
                  wrote last edited by
                  #32

                  @robotistry
                  @leeloo
                  so it's a parrot not because it's a matrix of probabilities, but because its hasn't experienced the real-world consequences of its words/actions and updated the probabilities based on those consequences?

                  0x00string@infosec.exchange0 robotistry@mstdn.caR 2 Replies Last reply
                  0
                  • kayohtie@blimps.xyzK kayohtie@blimps.xyz

                    @leeloo I hadn't thought about it as being something that takes magic away from folks like that. Honestly I always found it an accurate shortcut term for what's genuinely a fascinating but hilariously misused technology.

                    I think the worst part is then when folks hear "statistics" and go "See this is why it's safe to feed it raw data" and it's like oh my god NO.

                    calcifer@masto.hackers.townC This user is from outside of this forum
                    calcifer@masto.hackers.townC This user is from outside of this forum
                    calcifer@masto.hackers.town
                    wrote last edited by
                    #33

                    @KayOhtie @leeloo honestly it’s safe to feed a model pretty much anything

                    But where you direct the outputs and how they are acted upon can get incredibly dangerous amazingly quickly. There’s a common misbelief that if you’re careful about inputs, LLMs are safe; and that’s almost exactly backwards

                    kayohtie@blimps.xyzK 1 Reply Last reply
                    0
                    • calcifer@masto.hackers.townC calcifer@masto.hackers.town

                      @KayOhtie @leeloo honestly it’s safe to feed a model pretty much anything

                      But where you direct the outputs and how they are acted upon can get incredibly dangerous amazingly quickly. There’s a common misbelief that if you’re careful about inputs, LLMs are safe; and that’s almost exactly backwards

                      kayohtie@blimps.xyzK This user is from outside of this forum
                      kayohtie@blimps.xyzK This user is from outside of this forum
                      kayohtie@blimps.xyz
                      wrote last edited by
                      #34

                      @calcifer @leeloo I meant 'safe' not as in "data leakage", but "getting anything remotely accurate out of it"

                      1 Reply Last reply
                      0
                      • mudri@mathstodon.xyzM mudri@mathstodon.xyz

                        @lmorchard The ability to induce such a rule goes well beyond the OP's characterisation of what LLMs do.

                        calcifer@masto.hackers.townC This user is from outside of this forum
                        calcifer@masto.hackers.townC This user is from outside of this forum
                        calcifer@masto.hackers.town
                        wrote last edited by
                        #35

                        @mudri @lmorchard it’s not inductive at all though. It’s just parroting the patterns it sees in its training data. If it wasn’t common to see exchanges like that, the response would be utter nonsense.

                        People misunderstand what “training” is. It’s modeling the input. Humans develop the rules for how to model that input. Emergent properties of that process can easily *seem* like thinking or reason, but it’s an illusion.

                        1 Reply Last reply
                        0
                        • growlph@greywolf.socialG growlph@greywolf.social

                          @leeloo I feel like there are certain situations where a stochastic parrot is useful, many more situations where it is not, and alarmingly few people recognizing the difference.

                          calcifer@masto.hackers.townC This user is from outside of this forum
                          calcifer@masto.hackers.townC This user is from outside of this forum
                          calcifer@masto.hackers.town
                          wrote last edited by
                          #36

                          @growlph @leeloo this is the whole frustration I have with the polarization on the topic. There is genuinely utility. There’s also a very good argument that the utility doesn’t exceed the costs (socially, environmentally, etc).

                          But the hype is unreal and legitimately dangerous.

                          tal@mastodon.socialT 1 Reply Last reply
                          0
                          • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                            @leeloo on the flipside, I feel like some people use the term "stochastic parrot" or "it just completes the next token" to imply that "therefore it cannot be intelligent" - is that correct reasoning?

                            eestileib@tech.lgbtE This user is from outside of this forum
                            eestileib@tech.lgbtE This user is from outside of this forum
                            eestileib@tech.lgbt
                            wrote last edited by
                            #37

                            @wolf480pl @leeloo

                            Yes and I take that position.

                            1 Reply Last reply
                            0
                            • dragonfrog@mastodon.sdf.orgD dragonfrog@mastodon.sdf.org

                              @leeloo @wolf480pl @lmorchard I mean, I believe the human mind is the product of the physical human, largely of the brain (I don't believe in a non-physical soul), and it might indeed be basically an incredibly complex big bunch of matrix multiplications. And yeah I believe that's pretty magical.

                              lmorchard@masto.hackers.townL This user is from outside of this forum
                              lmorchard@masto.hackers.townL This user is from outside of this forum
                              lmorchard@masto.hackers.town
                              wrote last edited by
                              #38

                              @dragonfrog @leeloo @wolf480pl

                              "Imagine you have two machines. One you can open up and examine all of its workings, and if you give it every picture of a cat on the whole internet, it can reliably distinguish cats from non-cats. The other is a black box and it can also reliably distinguish cats from non-cats if you give it half a dozen pictures of cats, some apple sauce, and a hug. ... I am extremely confident in saying it doesn’t work the same way as the first one."

                              Link Preview Image
                              A.I. Isn't People

                              How many Reddit posts does it take to learn to read?

                              favicon

                              Today in Tabs (www.todayintabs.com)

                              dragonfrog@mastodon.sdf.orgD 1 Reply Last reply
                              0
                              • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                                @lmorchard @leeloo
                                I don't buy a general "no matrix multiplication will ever be intelligent".

                                jrdepriest@infosec.exchangeJ This user is from outside of this forum
                                jrdepriest@infosec.exchangeJ This user is from outside of this forum
                                jrdepriest@infosec.exchange
                                wrote last edited by
                                #39

                                @wolf480pl @lmorchard @leeloo you are allowed to believe that even if it is factually incorrect.

                                Link Preview Image
                                A non-anthropomorphized view of LLMs

                                In many discussions where questions of "alignment" or "AI safety" crop up, I am baffled by seriously intelligent people imbuing almost magic...

                                favicon

                                (addxorrol.blogspot.com)

                                Link Preview Image
                                Is language the same as intelligence? The AI industry desperately needs it to be

                                Neuroscience indicates language is distinct from thought, raising questions about whether AI large language models are a viable path to artificial general intelligence.

                                favicon

                                The Verge (www.theverge.com)

                                Just a moment...

                                favicon

                                (medium.com)

                                Link Preview Image
                                The LLMentalist Effect: how chat-based Large Language Models rep…

                                How to make better software with systems-thinking

                                favicon

                                Out of the Software Crisis (softwarecrisis.dev)

                                lmorchard@masto.hackers.townL 1 Reply Last reply
                                0
                                • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                                  @leeloo
                                  My point is that emergent properties can manifest even in systems ruled by very simple rules, and can be difficult to predict by just looking at the rules.

                                  And human intelligence, whatever it is, is likely an emergent property of human brain.

                                  Therefore, we cannot rule out that a similar emergent property will appear in artidicial systems that are not made of neurons without referring to how the neurons are arranged, and how the artificial systems are arranged.

                                  0x00string@infosec.exchange0 This user is from outside of this forum
                                  0x00string@infosec.exchange0 This user is from outside of this forum
                                  0x00string@infosec.exchange
                                  wrote last edited by
                                  #40

                                  @wolf480pl @leeloo dude its a spreadsheet

                                  1 Reply Last reply
                                  0
                                  • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                                    @robotistry
                                    @leeloo
                                    so it's a parrot not because it's a matrix of probabilities, but because its hasn't experienced the real-world consequences of its words/actions and updated the probabilities based on those consequences?

                                    0x00string@infosec.exchange0 This user is from outside of this forum
                                    0x00string@infosec.exchange0 This user is from outside of this forum
                                    0x00string@infosec.exchange
                                    wrote last edited by
                                    #41

                                    @wolf480pl @robotistry @leeloo spreadsheets cant have experiences, it doesnt update its probabilities, human beings spend insane amounts of money to generate the spreadsheets, nothing new comes out of them, have you uhh ever like looked into how this software works?

                                    1 Reply Last reply
                                    0
                                    • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                                      @lmorchard @leeloo
                                      I don't buy a general "no matrix multiplication will ever be intelligent".

                                      0x00string@infosec.exchange0 This user is from outside of this forum
                                      0x00string@infosec.exchange0 This user is from outside of this forum
                                      0x00string@infosec.exchange
                                      wrote last edited by
                                      #42

                                      @wolf480pl @lmorchard @leeloo praise be all glory to the llm

                                      lmorchard@masto.hackers.townL 1 Reply Last reply
                                      0
                                      • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                                        @lmorchard @leeloo
                                        I don't buy a general "no matrix multiplication will ever be intelligent".

                                        splendorr@mastodon.socialS This user is from outside of this forum
                                        splendorr@mastodon.socialS This user is from outside of this forum
                                        splendorr@mastodon.social
                                        wrote last edited by
                                        #43

                                        @wolf480pl @lmorchard @leeloo okay but that’s true. matrix multiplication will never be intelligent. the truth is neat!

                                        1 Reply Last reply
                                        0
                                        • clusterfcku@mastodon.socialC clusterfcku@mastodon.social

                                          @leeloo the flip side question about intelligence and LLMs is whether much of what we consider intelligence in humans is in fact just stochastic parrotting by humans.

                                          splendorr@mastodon.socialS This user is from outside of this forum
                                          splendorr@mastodon.socialS This user is from outside of this forum
                                          splendorr@mastodon.social
                                          wrote last edited by
                                          #44

                                          @clusterfcku @leeloo it’s not, and it sucks to suggest that

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups