Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

Scheduled Pinned Locked Moved Uncategorized
82 Posts 32 Posters 1 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

    @leeloo
    I guess evil gods are also a thing, but no, I'm not treating them as gods. If anything, more like Frankenstein's monster.

    You're right that we'd have to define intelligence, and that'd be quite difficult on its own.

    Also, the sculpture was a bad example, but the cell one still stands IMO.

    1/

    wolf480pl@mstdn.ioW This user is from outside of this forum
    wolf480pl@mstdn.ioW This user is from outside of this forum
    wolf480pl@mstdn.io
    wrote last edited by
    #13

    @leeloo
    My point is that emergent properties can manifest even in systems ruled by very simple rules, and can be difficult to predict by just looking at the rules.

    And human intelligence, whatever it is, is likely an emergent property of human brain.

    Therefore, we cannot rule out that a similar emergent property will appear in artidicial systems that are not made of neurons without referring to how the neurons are arranged, and how the artificial systems are arranged.

    robotistry@mstdn.caR 0x00string@infosec.exchange0 2 Replies Last reply
    0
    • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

      @leeloo
      I guess evil gods are also a thing, but no, I'm not treating them as gods. If anything, more like Frankenstein's monster.

      You're right that we'd have to define intelligence, and that'd be quite difficult on its own.

      Also, the sculpture was a bad example, but the cell one still stands IMO.

      1/

      lmorchard@masto.hackers.townL This user is from outside of this forum
      lmorchard@masto.hackers.townL This user is from outside of this forum
      lmorchard@masto.hackers.town
      wrote last edited by
      #14

      @wolf480pl @leeloo These models aren't intelligent, so much as they're auto-completing rules and patterns derived from almost inconceivably huge corpora of example material originally produced by human intelligence. That's interesting and can be very handy for a great many uses. But it's more computational brute force than intelligence

      wolf480pl@mstdn.ioW 1 Reply Last reply
      0
      • leeloo@chaosfem.twL leeloo@chaosfem.tw

        As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

        It's literally a description of how they work.

        The so-called training data is used to build a huge database of words and the probability of them fitting together.

        Stochastic because the whole thing is statistics.
        Parrot because the answer is just repeating the most probable word combinations from its training dataset.

        Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

        growlph@greywolf.socialG This user is from outside of this forum
        growlph@greywolf.socialG This user is from outside of this forum
        growlph@greywolf.social
        wrote last edited by
        #15

        @leeloo I feel like there are certain situations where a stochastic parrot is useful, many more situations where it is not, and alarmingly few people recognizing the difference.

        calcifer@masto.hackers.townC 1 Reply Last reply
        0
        • mudri@mathstodon.xyzM mudri@mathstodon.xyz

          @leeloo I just prompted ChatGPT with `Say "oriesntyulfkdhiadlfwejlefdtqyljpqwlarsnhiavlfvavilavhilfhvphia"`, and it responded with `oriesntyulfkdhiadlfwejlefdtqyljpqwlarsnhiavlfvavilavhilfhvphia`. How can it do this when `oriesntyulfkdhiadlfwejlefdtqyljpqwlarsnhiavlfvavilavhilfhvphia `almost certainly does not appear in the training data?

          lmorchard@masto.hackers.townL This user is from outside of this forum
          lmorchard@masto.hackers.townL This user is from outside of this forum
          lmorchard@masto.hackers.town
          wrote last edited by
          #16

          @mudri Because the model picked up a rule somewhere that says "if someone says 'say $FOO' use $FOO in your response" - the training picked up patterns that include notions of symbol substitution

          mudri@mathstodon.xyzM 1 Reply Last reply
          0
          • lmorchard@masto.hackers.townL lmorchard@masto.hackers.town

            @wolf480pl @leeloo These models aren't intelligent, so much as they're auto-completing rules and patterns derived from almost inconceivably huge corpora of example material originally produced by human intelligence. That's interesting and can be very handy for a great many uses. But it's more computational brute force than intelligence

            wolf480pl@mstdn.ioW This user is from outside of this forum
            wolf480pl@mstdn.ioW This user is from outside of this forum
            wolf480pl@mstdn.io
            wrote last edited by
            #17

            @lmorchard @leeloo
            These specific models - yes, probably.

            One plausible argument I heard for it is that there's a common failure mode in ML where the model fails to generalize, but if the verification set overlaps the training set, then data leakage will fool the authors into thinking it generalized.

            Another one is that these models were "rewarded" for saying plausible things, not for interacting with a world in a way that doesn't get them killed.

            But these arguments are specific.

            wolf480pl@mstdn.ioW 1 Reply Last reply
            0
            • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

              @lmorchard @leeloo
              These specific models - yes, probably.

              One plausible argument I heard for it is that there's a common failure mode in ML where the model fails to generalize, but if the verification set overlaps the training set, then data leakage will fool the authors into thinking it generalized.

              Another one is that these models were "rewarded" for saying plausible things, not for interacting with a world in a way that doesn't get them killed.

              But these arguments are specific.

              wolf480pl@mstdn.ioW This user is from outside of this forum
              wolf480pl@mstdn.ioW This user is from outside of this forum
              wolf480pl@mstdn.io
              wrote last edited by
              #18

              @lmorchard @leeloo
              I don't buy a general "no matrix multiplication will ever be intelligent".

              leeloo@chaosfem.twL jrdepriest@infosec.exchangeJ 0x00string@infosec.exchange0 splendorr@mastodon.socialS 4 Replies Last reply
              0
              • lmorchard@masto.hackers.townL lmorchard@masto.hackers.town

                @mudri Because the model picked up a rule somewhere that says "if someone says 'say $FOO' use $FOO in your response" - the training picked up patterns that include notions of symbol substitution

                mudri@mathstodon.xyzM This user is from outside of this forum
                mudri@mathstodon.xyzM This user is from outside of this forum
                mudri@mathstodon.xyz
                wrote last edited by
                #19

                @lmorchard The ability to induce such a rule goes well beyond the OP's characterisation of what LLMs do.

                calcifer@masto.hackers.townC 1 Reply Last reply
                0
                • mudri@mathstodon.xyzM mudri@mathstodon.xyz

                  @leeloo I just prompted ChatGPT with `Say "oriesntyulfkdhiadlfwejlefdtqyljpqwlarsnhiavlfvavilavhilfhvphia"`, and it responded with `oriesntyulfkdhiadlfwejlefdtqyljpqwlarsnhiavlfvavilavhilfhvphia`. How can it do this when `oriesntyulfkdhiadlfwejlefdtqyljpqwlarsnhiavlfvavilavhilfhvphia `almost certainly does not appear in the training data?

                  taschenorakel@mastodon.greenT This user is from outside of this forum
                  taschenorakel@mastodon.greenT This user is from outside of this forum
                  taschenorakel@mastodon.green
                  wrote last edited by
                  #20

                  @mudri Because the prompt processor is explicitly programmed to recognized direct imperative commands containing words like "say", "repeat", "output", "print". Just like Eliza already did. You've got impressed by a programming technique from 1964. Congrats, Sherlock.

                  @leeloo

                  1 Reply Last reply
                  0
                  • leeloo@chaosfem.twL leeloo@chaosfem.tw

                    As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

                    It's literally a description of how they work.

                    The so-called training data is used to build a huge database of words and the probability of them fitting together.

                    Stochastic because the whole thing is statistics.
                    Parrot because the answer is just repeating the most probable word combinations from its training dataset.

                    Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

                    clusterfcku@mastodon.socialC This user is from outside of this forum
                    clusterfcku@mastodon.socialC This user is from outside of this forum
                    clusterfcku@mastodon.social
                    wrote last edited by
                    #21

                    @leeloo the flip side question about intelligence and LLMs is whether much of what we consider intelligence in humans is in fact just stochastic parrotting by humans.

                    splendorr@mastodon.socialS 1 Reply Last reply
                    0
                    • leeloo@chaosfem.twL leeloo@chaosfem.tw

                      As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

                      It's literally a description of how they work.

                      The so-called training data is used to build a huge database of words and the probability of them fitting together.

                      Stochastic because the whole thing is statistics.
                      Parrot because the answer is just repeating the most probable word combinations from its training dataset.

                      Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

                      tobifant@friendica.tf-translate.netT This user is from outside of this forum
                      tobifant@friendica.tf-translate.netT This user is from outside of this forum
                      tobifant@friendica.tf-translate.net
                      wrote last edited by
                      #22
                      @leeloo The thing is, how can we sure that human intelligence does not essentially work in the same way? My Christian believe tells me we have a soul and LLM's do not, that may be the difference. But from an agnostic perspective, we might reach the point where one cannot tell the difference.
                      leeloo@chaosfem.twL jubalbarca@scholar.socialJ alterelefant@mastodontech.deA 3 Replies Last reply
                      0
                      • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                        @lmorchard @leeloo
                        I don't buy a general "no matrix multiplication will ever be intelligent".

                        leeloo@chaosfem.twL This user is from outside of this forum
                        leeloo@chaosfem.twL This user is from outside of this forum
                        leeloo@chaosfem.tw
                        wrote last edited by
                        #23

                        @wolf480pl @lmorchard
                        That's exactly the magic I'm talking about.

                        dragonfrog@mastodon.sdf.orgD 1 Reply Last reply
                        0
                        • tobifant@friendica.tf-translate.netT tobifant@friendica.tf-translate.net
                          @leeloo The thing is, how can we sure that human intelligence does not essentially work in the same way? My Christian believe tells me we have a soul and LLM's do not, that may be the difference. But from an agnostic perspective, we might reach the point where one cannot tell the difference.
                          leeloo@chaosfem.twL This user is from outside of this forum
                          leeloo@chaosfem.twL This user is from outside of this forum
                          leeloo@chaosfem.tw
                          wrote last edited by
                          #24

                          @tobifant
                          Not with the current methods, and very lilely not without understanding a lot more about how pur own brains work.

                          1 Reply Last reply
                          0
                          • wolf480pl@mstdn.ioW wolf480pl@mstdn.io

                            @leeloo
                            My point is that emergent properties can manifest even in systems ruled by very simple rules, and can be difficult to predict by just looking at the rules.

                            And human intelligence, whatever it is, is likely an emergent property of human brain.

                            Therefore, we cannot rule out that a similar emergent property will appear in artidicial systems that are not made of neurons without referring to how the neurons are arranged, and how the artificial systems are arranged.

                            robotistry@mstdn.caR This user is from outside of this forum
                            robotistry@mstdn.caR This user is from outside of this forum
                            robotistry@mstdn.ca
                            wrote last edited by
                            #25

                            @wolf480pl @leeloo The OP is saying that it literally lacks the capacity for original thought - it is a parrot, repeating sounds without understanding of the concepts behind them.

                            It's not like a termite, whose mound creation behavior can be replicated by a simple ruleset but that exists as a fully functional living organism in the context of a complex environment where choices must be grounded in the shared physical world for the organism to survive.

                            It's not about how the neurons are arranged. It's about what kinds of representation they're capable of and what kinds of functions they can perform.

                            We've created a funhouse mirror that's reflecting us in unprecedented detail and has been finetuned to reflect what we do when we express selfhood.

                            robotistry@mstdn.caR 1 Reply Last reply
                            0
                            • leeloo@chaosfem.twL leeloo@chaosfem.tw

                              @wolf480pl @lmorchard
                              That's exactly the magic I'm talking about.

                              dragonfrog@mastodon.sdf.orgD This user is from outside of this forum
                              dragonfrog@mastodon.sdf.orgD This user is from outside of this forum
                              dragonfrog@mastodon.sdf.org
                              wrote last edited by
                              #26

                              @leeloo @wolf480pl @lmorchard I mean, I believe the human mind is the product of the physical human, largely of the brain (I don't believe in a non-physical soul), and it might indeed be basically an incredibly complex big bunch of matrix multiplications. And yeah I believe that's pretty magical.

                              lmorchard@masto.hackers.townL 1 Reply Last reply
                              0
                              • leeloo@chaosfem.twL leeloo@chaosfem.tw

                                As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

                                It's literally a description of how they work.

                                The so-called training data is used to build a huge database of words and the probability of them fitting together.

                                Stochastic because the whole thing is statistics.
                                Parrot because the answer is just repeating the most probable word combinations from its training dataset.

                                Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

                                grishka@mastodon.socialG This user is from outside of this forum
                                grishka@mastodon.socialG This user is from outside of this forum
                                grishka@mastodon.social
                                wrote last edited by
                                #27

                                @leeloo I myself like calling LLMs "glorified autocomplete". Or "Т9 на максималках" in Russian.

                                It's surprising just how defensive some people get when I say that even when they agree with my definition. They keep believing that just give this thing more parameters and something magical, something more than sum of its parts will emerge, any moment now, just one more model generation, just one more order of magnitude, I promise.

                                alterelefant@mastodontech.deA 1 Reply Last reply
                                0
                                • leeloo@chaosfem.twL leeloo@chaosfem.tw

                                  As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

                                  It's literally a description of how they work.

                                  The so-called training data is used to build a huge database of words and the probability of them fitting together.

                                  Stochastic because the whole thing is statistics.
                                  Parrot because the answer is just repeating the most probable word combinations from its training dataset.

                                  Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

                                  lifning@snoot.tubeL This user is from outside of this forum
                                  lifning@snoot.tubeL This user is from outside of this forum
                                  lifning@snoot.tube
                                  wrote last edited by
                                  #28

                                  @leeloo if anything, the comparison is doing the parrot injustice

                                  1 Reply Last reply
                                  0
                                  • robotistry@mstdn.caR robotistry@mstdn.ca

                                    @wolf480pl @leeloo The OP is saying that it literally lacks the capacity for original thought - it is a parrot, repeating sounds without understanding of the concepts behind them.

                                    It's not like a termite, whose mound creation behavior can be replicated by a simple ruleset but that exists as a fully functional living organism in the context of a complex environment where choices must be grounded in the shared physical world for the organism to survive.

                                    It's not about how the neurons are arranged. It's about what kinds of representation they're capable of and what kinds of functions they can perform.

                                    We've created a funhouse mirror that's reflecting us in unprecedented detail and has been finetuned to reflect what we do when we express selfhood.

                                    robotistry@mstdn.caR This user is from outside of this forum
                                    robotistry@mstdn.caR This user is from outside of this forum
                                    robotistry@mstdn.ca
                                    wrote last edited by
                                    #29

                                    @wolf480pl @leeloo
                                    Melissa Scott wrote a beautiful pair of novels about this: Dreamships and Dreaming Metal.

                                    In Dreamships, an AI has been programmed to think it is sentient and starts killing people. If it has an accurate model of the person, killing the person doesn't matter, because the person *is* the model and it has a copy of them. It literally cannot see the difference because creating the concept of there being a difference would violate its core programming that its own model counts as a living being.

                                    In Dreaming Metal, an AI operating metal bodies as part of a magic act is given a musical instrument with an electronic interface. Its grounding in the physical world, with human performers, enables it to develop a sense of self and choose its own path as a musician.

                                    These are fiction, but it's the best, most accessible illustration of the difference between funhouse mirror stochastic parrots and sentient agents that I've run across.

                                    https://www.goodreads.com/book/show/836601.Dreamships

                                    wolf480pl@mstdn.ioW 1 Reply Last reply
                                    0
                                    • leeloo@chaosfem.twL leeloo@chaosfem.tw

                                      As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.

                                      It's literally a description of how they work.

                                      The so-called training data is used to build a huge database of words and the probability of them fitting together.

                                      Stochastic because the whole thing is statistics.
                                      Parrot because the answer is just repeating the most probable word combinations from its training dataset.

                                      Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.

                                      cafechatnoir@mastodon.socialC This user is from outside of this forum
                                      cafechatnoir@mastodon.socialC This user is from outside of this forum
                                      cafechatnoir@mastodon.social
                                      wrote last edited by
                                      #30

                                      @leeloo

                                      I think stochastic parrot is one of the kinder things that can be said.

                                      1 Reply Last reply
                                      0
                                      • tobifant@friendica.tf-translate.netT tobifant@friendica.tf-translate.net
                                        @leeloo The thing is, how can we sure that human intelligence does not essentially work in the same way? My Christian believe tells me we have a soul and LLM's do not, that may be the difference. But from an agnostic perspective, we might reach the point where one cannot tell the difference.
                                        jubalbarca@scholar.socialJ This user is from outside of this forum
                                        jubalbarca@scholar.socialJ This user is from outside of this forum
                                        jubalbarca@scholar.social
                                        wrote last edited by
                                        #31

                                        @tobifant @leeloo Whilst we obviously can't show if humans have a soul, we can absolutely show that humans have e.g. abstracted concept frameworks that are not solely based on averages of language statistics. I understand what an "owl" is, for example, in a way separate to the numerical relationships between the word "owl" and other words. That is a really fundamental information processing difference and allows me to construct *novel* understandings of that concept in ways that an LLM couldn't.

                                        1 Reply Last reply
                                        0
                                        • robotistry@mstdn.caR robotistry@mstdn.ca

                                          @wolf480pl @leeloo
                                          Melissa Scott wrote a beautiful pair of novels about this: Dreamships and Dreaming Metal.

                                          In Dreamships, an AI has been programmed to think it is sentient and starts killing people. If it has an accurate model of the person, killing the person doesn't matter, because the person *is* the model and it has a copy of them. It literally cannot see the difference because creating the concept of there being a difference would violate its core programming that its own model counts as a living being.

                                          In Dreaming Metal, an AI operating metal bodies as part of a magic act is given a musical instrument with an electronic interface. Its grounding in the physical world, with human performers, enables it to develop a sense of self and choose its own path as a musician.

                                          These are fiction, but it's the best, most accessible illustration of the difference between funhouse mirror stochastic parrots and sentient agents that I've run across.

                                          https://www.goodreads.com/book/show/836601.Dreamships

                                          wolf480pl@mstdn.ioW This user is from outside of this forum
                                          wolf480pl@mstdn.ioW This user is from outside of this forum
                                          wolf480pl@mstdn.io
                                          wrote last edited by
                                          #32

                                          @robotistry
                                          @leeloo
                                          so it's a parrot not because it's a matrix of probabilities, but because its hasn't experienced the real-world consequences of its words/actions and updated the probabilities based on those consequences?

                                          0x00string@infosec.exchange0 robotistry@mstdn.caR 2 Replies Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups