As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.
-
@lmorchard @leeloo
I don't buy a general "no matrix multiplication will ever be intelligent".@wolf480pl @lmorchard @leeloo praise be all glory to the llm
-
@lmorchard @leeloo
I don't buy a general "no matrix multiplication will ever be intelligent".@wolf480pl @lmorchard @leeloo okay but that’s true. matrix multiplication will never be intelligent. the truth is neat!
-
@leeloo the flip side question about intelligence and LLMs is whether much of what we consider intelligence in humans is in fact just stochastic parrotting by humans.
@clusterfcku @leeloo it’s not, and it sucks to suggest that
-
@wolf480pl @lmorchard @leeloo praise be all glory to the llm
-
@wolf480pl @lmorchard @leeloo you are allowed to believe that even if it is factually incorrect.
A non-anthropomorphized view of LLMs
In many discussions where questions of "alignment" or "AI safety" crop up, I am baffled by seriously intelligent people imbuing almost magic...
(addxorrol.blogspot.com)
Is language the same as intelligence? The AI industry desperately needs it to be
Neuroscience indicates language is distinct from thought, raising questions about whether AI large language models are a viable path to artificial general intelligence.
The Verge (www.theverge.com)
The LLMentalist Effect: how chat-based Large Language Models rep…
How to make better software with systems-thinking
Out of the Software Crisis (softwarecrisis.dev)
@jrdepriest @wolf480pl @leeloo I'm confused... those links basically say what I said. (i.e. the "intelligence" is second-hand) That's... incorrect?
-
@jrdepriest @wolf480pl @leeloo I'm confused... those links basically say what I said. (i.e. the "intelligence" is second-hand) That's... incorrect?
LLM based genAI can never be "intelligent". They can spit out language that looks like intelligence but there is no thinking, no inner life, no thoughts, just math. And this is not how the human brain works.
Also, we know the brain is not a computer.
-
@dragonfrog @leeloo @wolf480pl
"Imagine you have two machines. One you can open up and examine all of its workings, and if you give it every picture of a cat on the whole internet, it can reliably distinguish cats from non-cats. The other is a black box and it can also reliably distinguish cats from non-cats if you give it half a dozen pictures of cats, some apple sauce, and a hug. ... I am extremely confident in saying it doesn’t work the same way as the first one."
A.I. Isn't People
How many Reddit posts does it take to learn to read?
Today in Tabs (www.todayintabs.com)
@lmorchard @leeloo @wolf480pl good grief now I have to sound like Sam friggin Altman, and there is clearly something very wrong with that man.
But your description ignores that humans need a solid 6 months of "training data" to get object permanence, never mind the concept of categories or species of animals, never mind understanding the category differences between cats and foxes well enough to reliably tell one from the other.
-
As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.
It's literally a description of how they work.
The so-called training data is used to build a huge database of words and the probability of them fitting together.
Stochastic because the whole thing is statistics.
Parrot because the answer is just repeating the most probable word combinations from its training dataset.Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.
-
@lmorchard @leeloo @wolf480pl good grief now I have to sound like Sam friggin Altman, and there is clearly something very wrong with that man.
But your description ignores that humans need a solid 6 months of "training data" to get object permanence, never mind the concept of categories or species of animals, never mind understanding the category differences between cats and foxes well enough to reliably tell one from the other.
@lmorchard @leeloo @wolf480pl I guess part of it is maybe that I don't think intelligence is some exclusively human thing. LLMs clearly aren't human-like intelligent. I'm personally confident they're not as intelligent as any primate.
But are they as intelligent as a shrimp? I think they've got to be more intelligent than a mosquito.
I wouldn't turn to a shrimp for advice but they're not *without* intelligence.
-
@lmorchard @leeloo @wolf480pl I guess part of it is maybe that I don't think intelligence is some exclusively human thing. LLMs clearly aren't human-like intelligent. I'm personally confident they're not as intelligent as any primate.
But are they as intelligent as a shrimp? I think they've got to be more intelligent than a mosquito.
I wouldn't turn to a shrimp for advice but they're not *without* intelligence.
@dragonfrog @lmorchard @leeloo @wolf480pl
Are the images reflected in a distorted mirror the product of intelligence (of the mirror)?
They are coherent, a literal transform of the input images, reflected and produce a recognizable, if distorted and changed version.
A traditional function output. Let's add some noise to make it non-deterministic, a wind blowing through that minutely distorts the surface.
Intelligible output following from the input, but the mirror itself isn't intelligent.
-
@dragonfrog @lmorchard @leeloo @wolf480pl
Are the images reflected in a distorted mirror the product of intelligence (of the mirror)?
They are coherent, a literal transform of the input images, reflected and produce a recognizable, if distorted and changed version.
A traditional function output. Let's add some noise to make it non-deterministic, a wind blowing through that minutely distorts the surface.
Intelligible output following from the input, but the mirror itself isn't intelligent.
@dragonfrog @lmorchard @leeloo @wolf480pl
The intelligence apparently making the meaning is pre-encoded in the input. Likewise, the vector math is extracting and exposing structure, encoded in language, put there originally by the intelligent humans.
There is no world model or understanding. That's why counting the "r" in strawberry or simply counting to 200 is so challenging.
The behavior can reasonably be called intelligent, but it's due to borrowed, reformulated, extracted intelligence
-
As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.
It's literally a description of how they work.
The so-called training data is used to build a huge database of words and the probability of them fitting together.
Stochastic because the whole thing is statistics.
Parrot because the answer is just repeating the most probable word combinations from its training dataset.Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.
@leeloo As a side note, I sometimes worry about how much parroting happens in academia among humans even before/without LLMs, where people repeat things without understanding what they’re talking about. I guess at least for students, it sometimes is about learning to talk the talk, and then gradually developing more understanding and genuine thinking around topics. At least we humans are capable of developing that understanding if we bother to try.
-
@leeloo The thing is, how can we sure that human intelligence does not essentially work in the same way? My Christian believe tells me we have a soul and LLM's do not, that may be the difference. But from an agnostic perspective, we might reach the point where one cannot tell the difference.
@tobifant
A LLM is not able to reason. It can fool you into believing it is intelligent and self aware, where in fact it just parrots the patterns it has stored. These patterns are however very human-like as they are the result of training on texts written by actual humans.The fun part starts now where the entire internet got flooded by #ai generated content. All of this will be the training set for the next generation of LLM's. What could possibly go wrong?
@leeloo -
@leeloo I myself like calling LLMs "glorified autocomplete". Or "Т9 на максималках" in Russian.
It's surprising just how defensive some people get when I say that even when they agree with my definition. They keep believing that just give this thing more parameters and something magical, something more than sum of its parts will emerge, any moment now, just one more model generation, just one more order of magnitude, I promise.
@grishka
The fun part is that the next generation will have the current state of the internet as its training set. An internet that is flooded by #ai generated content.The biggest issue those ai companies face at the moment is how to only ingest human generated content and filter out as much as possible of all of the ai generated crap that is out there.
Good luck with that.
@leeloo -
-
R relay@relay.an.exchange shared this topic
-
As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.
It's literally a description of how they work.
The so-called training data is used to build a huge database of words and the probability of them fitting together.
Stochastic because the whole thing is statistics.
Parrot because the answer is just repeating the most probable word combinations from its training dataset.Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.
@leeloo If I want to disparage, I say "LLMs are just a word list with a randomizer". It's slightly less accurate, because it's a very specific kind of word list and a likewise specific kind of randomizer, but it gets the cultists all riled up. So that's cool.
-
As a software developer who took an elective in neural networks - when people call LLMs stochastic parrots, that's not criticism of their results.
It's literally a description of how they work.
The so-called training data is used to build a huge database of words and the probability of them fitting together.
Stochastic because the whole thing is statistics.
Parrot because the answer is just repeating the most probable word combinations from its training dataset.Calling an LLM a stochastic parrot is lile calling a car a motorised vehicle with wheels. It doesn't say anything about cars being good or bad. It does, however, take away the magic. So if you feel a need to defend AI when you hear the term stochastic parrot, consider that you may have elevated them to a god-like status, and that's why you go on the defense when the magic is dispelled.
nope. What you describe as "stocastical parrot" is Markov, Hidden Markov Model (HMM) , not a VLLM.
You can find an HMM in your mobile phone, AKA T9, AKA "keyboard suggestions".
-
R relay@relay.mycrowd.ca shared this topic
-
nope. What you describe as "stocastical parrot" is Markov, Hidden Markov Model (HMM) , not a VLLM.
You can find an HMM in your mobile phone, AKA T9, AKA "keyboard suggestions".
@uriel
What part exactly are you saying nope to.Dispelling the magic and god-like status or some specific detail?
-
@uriel
What part exactly are you saying nope to.Dispelling the magic and god-like status or some specific detail?
nope to the bunch of bullshit you wrote under the assumption a VLLM is a Hidden Markov Model , aka "stochastic parrot".
-
@calcifer
> But the hype is unreal and legitimately dangerous.I blame Sam Altman for that 100%