Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing.

My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing.

Scheduled Pinned Locked Moved Uncategorized
40 Posts 31 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

    My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

    LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

    In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

    But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

    If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

    evildrganymede@wargamers.socialE This user is from outside of this forum
    evildrganymede@wargamers.socialE This user is from outside of this forum
    evildrganymede@wargamers.social
    wrote last edited by
    #16

    @EmilyEnough I think you're absolutely correct on this. Yet another reason why we need to find a way to irrevocably destroy this abomination.

    But also it's not just the style of "communication" that these algorithms are pretending to do, it's that you cannot trust that their output is even correct because they have no understanding of what they are "saying". They could be "hallucinating" complete nonsense but they'll output it in an authoritative way and may even make up references that don't exist. They're 100% bullshit generators (it's even been scientifically proven).

    1 Reply Last reply
    0
    • uriel@x.keinpfusch.netU This user is from outside of this forum
      uriel@x.keinpfusch.netU This user is from outside of this forum
      uriel@x.keinpfusch.net
      wrote last edited by
      #17

      @EmilyEnough

      My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine

      Also known as “training”. When people are trained in art, they don’t reinvent art from scratch. This is why you can’t really sue an LLM for plagiarism: you can’t even identify specific victims in the first place.

      and disaster for the environment,

      Nope. The whole IT sector uses about 3–5% of global electricity, so poor home insulation is a much bigger problem overall.

      is that they introduce so much unpredictability into computing.

      We call it a statistical method, or more precisely a stochastic system. Because, to a large extent, human behaviour itself can be modelled as a stochastic process.

      If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering.

      The problems you face when communicating with LLMs are the same ones you face when communicating with people, because statistically speaking an LLM mimics how people communicate.

      This is why computer‑mediated communication was used before, and is still used, when computers were not trying to mimic humans.

      The core issue is that mimicking humans reproduces the same communication problems people already have with one another; and the “unpredictability” of the other party is nothing new in human interaction.

      LLMs mimic humans, so the problems you encounter with LLMs are the same problems you encounter with humans. The point is that you consider it normal when you face exactly the same issues with other people.

      magitweeter@mastodon.socialM pglpm@c.imP greenskyoverme@ohai.socialG 3 Replies Last reply
      0
      • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

        My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

        LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

        In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

        But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

        If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

        monkee@chaos.socialM This user is from outside of this forum
        monkee@chaos.socialM This user is from outside of this forum
        monkee@chaos.social
        wrote last edited by
        #18

        @EmilyEnough

        "They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem."

        OMG - That's perfect. Maybe also explains why everyone loves them that much. 🤨

        1 Reply Last reply
        0
        • uriel@x.keinpfusch.netU uriel@x.keinpfusch.net

          @EmilyEnough

          My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine

          Also known as “training”. When people are trained in art, they don’t reinvent art from scratch. This is why you can’t really sue an LLM for plagiarism: you can’t even identify specific victims in the first place.

          and disaster for the environment,

          Nope. The whole IT sector uses about 3–5% of global electricity, so poor home insulation is a much bigger problem overall.

          is that they introduce so much unpredictability into computing.

          We call it a statistical method, or more precisely a stochastic system. Because, to a large extent, human behaviour itself can be modelled as a stochastic process.

          If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering.

          The problems you face when communicating with LLMs are the same ones you face when communicating with people, because statistically speaking an LLM mimics how people communicate.

          This is why computer‑mediated communication was used before, and is still used, when computers were not trying to mimic humans.

          The core issue is that mimicking humans reproduces the same communication problems people already have with one another; and the “unpredictability” of the other party is nothing new in human interaction.

          LLMs mimic humans, so the problems you encounter with LLMs are the same problems you encounter with humans. The point is that you consider it normal when you face exactly the same issues with other people.

          magitweeter@mastodon.socialM This user is from outside of this forum
          magitweeter@mastodon.socialM This user is from outside of this forum
          magitweeter@mastodon.social
          wrote last edited by
          #19

          @uriel Way to miss the point

          @EmilyEnough

          uriel@x.keinpfusch.netU 1 Reply Last reply
          0
          • magitweeter@mastodon.socialM magitweeter@mastodon.social

            @uriel Way to miss the point

            @EmilyEnough

            uriel@x.keinpfusch.netU This user is from outside of this forum
            uriel@x.keinpfusch.netU This user is from outside of this forum
            uriel@x.keinpfusch.net
            wrote last edited by
            #20

            @magitweeter @EmilyEnough

            Way to miss the point

            whatever.

            1 Reply Last reply
            0
            • uriel@x.keinpfusch.netU uriel@x.keinpfusch.net

              @EmilyEnough

              My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine

              Also known as “training”. When people are trained in art, they don’t reinvent art from scratch. This is why you can’t really sue an LLM for plagiarism: you can’t even identify specific victims in the first place.

              and disaster for the environment,

              Nope. The whole IT sector uses about 3–5% of global electricity, so poor home insulation is a much bigger problem overall.

              is that they introduce so much unpredictability into computing.

              We call it a statistical method, or more precisely a stochastic system. Because, to a large extent, human behaviour itself can be modelled as a stochastic process.

              If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering.

              The problems you face when communicating with LLMs are the same ones you face when communicating with people, because statistically speaking an LLM mimics how people communicate.

              This is why computer‑mediated communication was used before, and is still used, when computers were not trying to mimic humans.

              The core issue is that mimicking humans reproduces the same communication problems people already have with one another; and the “unpredictability” of the other party is nothing new in human interaction.

              LLMs mimic humans, so the problems you encounter with LLMs are the same problems you encounter with humans. The point is that you consider it normal when you face exactly the same issues with other people.

              pglpm@c.imP This user is from outside of this forum
              pglpm@c.imP This user is from outside of this forum
              pglpm@c.im
              wrote last edited by
              #21

              @uriel @EmilyEnough

              > Nope. The whole IT sector uses about 3–5% of global electricity, so poor home insulation is a much bigger problem overall.

              Source?

              > We call it a statistical method, or more precisely a stochastic system. Because, to a large extent, human behaviour itself can be modelled as a stochastic process.

              Source? In fact this is false. Human behaviour includes more than a stochastic process, even though it may adopt stochastic heuristics to speed up some computational parts. This is also why LLMs are technically speaking *not* AI. An AI includes, as human reasoning does, an internal world model and the basic set of Boolean probability-logic rules. See for instance Russell & Norvig's *Artificial Intelligence: A Modern Approach* (http://aima.cs.berkeley.edu/global-index.html), or Pearl's older *Probabilistic Reasoning in Intelligent Systems* (https://doi.org/10.1016/C2009-0-27609-4). LLMs are, instead, just Markov chains (https://doi.org/10.48550/arXiv.2410.02724). A modern robot vacuum cleaner is more "AI" than an LLM.

              This is also the reason why the larger the software project you apply an LLM to, the more likely the failure. Such kind of application requires larger and larger string correlations, which are therefore more and more uncertain and fault-prone, and these faults are therefore also more difficult to spot. Such kind of applications may also require new or innovative kinds of solution, which again are less likely to be stumbled upon by an LLM.

              > The problems you face when communicating with LLMs are the same ones you face when communicating with people, because statistically speaking an LLM mimics how people communicate.

              No, because humans, and also *proper AI*, have a "logic engine" underneath. It may require some effort to bring the logic engine to the fore instead of poor heuristics, but it can be done (related: Kahneman's *Thinking, Fast and Slow*, and the research cited there). With LLM it can't be done because there's no logic engine at all there.

              uriel@x.keinpfusch.netU 1 Reply Last reply
              0
              • pglpm@c.imP pglpm@c.im

                @uriel @EmilyEnough

                > Nope. The whole IT sector uses about 3–5% of global electricity, so poor home insulation is a much bigger problem overall.

                Source?

                > We call it a statistical method, or more precisely a stochastic system. Because, to a large extent, human behaviour itself can be modelled as a stochastic process.

                Source? In fact this is false. Human behaviour includes more than a stochastic process, even though it may adopt stochastic heuristics to speed up some computational parts. This is also why LLMs are technically speaking *not* AI. An AI includes, as human reasoning does, an internal world model and the basic set of Boolean probability-logic rules. See for instance Russell & Norvig's *Artificial Intelligence: A Modern Approach* (http://aima.cs.berkeley.edu/global-index.html), or Pearl's older *Probabilistic Reasoning in Intelligent Systems* (https://doi.org/10.1016/C2009-0-27609-4). LLMs are, instead, just Markov chains (https://doi.org/10.48550/arXiv.2410.02724). A modern robot vacuum cleaner is more "AI" than an LLM.

                This is also the reason why the larger the software project you apply an LLM to, the more likely the failure. Such kind of application requires larger and larger string correlations, which are therefore more and more uncertain and fault-prone, and these faults are therefore also more difficult to spot. Such kind of applications may also require new or innovative kinds of solution, which again are less likely to be stumbled upon by an LLM.

                > The problems you face when communicating with LLMs are the same ones you face when communicating with people, because statistically speaking an LLM mimics how people communicate.

                No, because humans, and also *proper AI*, have a "logic engine" underneath. It may require some effort to bring the logic engine to the fore instead of poor heuristics, but it can be done (related: Kahneman's *Thinking, Fast and Slow*, and the research cited there). With LLM it can't be done because there's no logic engine at all there.

                uriel@x.keinpfusch.netU This user is from outside of this forum
                uriel@x.keinpfusch.netU This user is from outside of this forum
                uriel@x.keinpfusch.net
                wrote last edited by
                #22

                @pglpm @EmilyEnough

                Source?

                Who Am I, your secretary? Just google.

                here is my answer, complete.

                Das Böse Büro

                blog, personale, pensieri, riflessioni, sfoghi , uriel, fanelli, loweel

                favicon

                Das Böse Büro (keinpfusch.net)

                pglpm@c.imP 1 Reply Last reply
                0
                • uriel@x.keinpfusch.netU uriel@x.keinpfusch.net

                  @pglpm @EmilyEnough

                  Source?

                  Who Am I, your secretary? Just google.

                  here is my answer, complete.

                  Das Böse Büro

                  blog, personale, pensieri, riflessioni, sfoghi , uriel, fanelli, loweel

                  favicon

                  Das Böse Büro (keinpfusch.net)

                  pglpm@c.imP This user is from outside of this forum
                  pglpm@c.imP This user is from outside of this forum
                  pglpm@c.im
                  wrote last edited by
                  #23

                  @uriel @EmilyEnough
                  No, you're the one making the claim, so the onus is on you to give evidence.

                  uriel@x.keinpfusch.netU 1 Reply Last reply
                  0
                  • pglpm@c.imP pglpm@c.im

                    @uriel @EmilyEnough
                    No, you're the one making the claim, so the onus is on you to give evidence.

                    uriel@x.keinpfusch.netU This user is from outside of this forum
                    uriel@x.keinpfusch.netU This user is from outside of this forum
                    uriel@x.keinpfusch.net
                    wrote last edited by
                    #24

                    @pglpm @EmilyEnough

                    ok, since you aren't able to, let me google for sources:

                    Link Preview Image
                    We did the math on AI’s energy footprint. Here’s the story you haven’t heard.

                    The emissions from individual AI text, image, and video queries seem small—until you add up what the industry isn’t tracking and consider where it’s heading next.

                    favicon

                    MIT Technology Review (www.technologyreview.com)

                    MIT says , 4.4%.

                    Arxiv is so full of shit, I don't even care. WARNING: next time you ask me to google something for you, since you are too stupid for , you must pay me.

                    Link Preview Image
                    pglpm@c.imP 1 Reply Last reply
                    0
                    • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

                      My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                      LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                      In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                      But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                      If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                      jesterchen@social.tchncs.deJ This user is from outside of this forum
                      jesterchen@social.tchncs.deJ This user is from outside of this forum
                      jesterchen@social.tchncs.de
                      wrote last edited by
                      #25

                      @EmilyEnough 🏆🏆🏆

                      1 Reply Last reply
                      0
                      • uriel@x.keinpfusch.netU uriel@x.keinpfusch.net

                        @EmilyEnough

                        My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine

                        Also known as “training”. When people are trained in art, they don’t reinvent art from scratch. This is why you can’t really sue an LLM for plagiarism: you can’t even identify specific victims in the first place.

                        and disaster for the environment,

                        Nope. The whole IT sector uses about 3–5% of global electricity, so poor home insulation is a much bigger problem overall.

                        is that they introduce so much unpredictability into computing.

                        We call it a statistical method, or more precisely a stochastic system. Because, to a large extent, human behaviour itself can be modelled as a stochastic process.

                        If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering.

                        The problems you face when communicating with LLMs are the same ones you face when communicating with people, because statistically speaking an LLM mimics how people communicate.

                        This is why computer‑mediated communication was used before, and is still used, when computers were not trying to mimic humans.

                        The core issue is that mimicking humans reproduces the same communication problems people already have with one another; and the “unpredictability” of the other party is nothing new in human interaction.

                        LLMs mimic humans, so the problems you encounter with LLMs are the same problems you encounter with humans. The point is that you consider it normal when you face exactly the same issues with other people.

                        greenskyoverme@ohai.socialG This user is from outside of this forum
                        greenskyoverme@ohai.socialG This user is from outside of this forum
                        greenskyoverme@ohai.social
                        wrote last edited by
                        #26

                        @uriel

                        No.

                        @EmilyEnough

                        uriel@x.keinpfusch.netU 1 Reply Last reply
                        0
                        • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

                          My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                          LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                          In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                          But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                          If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                          lettosprey@tech.lgbtL This user is from outside of this forum
                          lettosprey@tech.lgbtL This user is from outside of this forum
                          lettosprey@tech.lgbt
                          wrote last edited by
                          #27

                          @EmilyEnough There are so many "My biggest problem with LLM, even if it wasn't for <list of other big problems>", there should be collection of them somewhere.

                          But, yes, this bit bugs (pun intended) me and worries me. I'm more and more falling for BEAM family languages (Erlang, Elixir and Gleam) because of how they are designed to be as predictable as possible.

                          It may not be too odd that I see a lot less AI push in that ecosystem compared to other ones.

                          1 Reply Last reply
                          0
                          • greenskyoverme@ohai.socialG greenskyoverme@ohai.social

                            @uriel

                            No.

                            @EmilyEnough

                            uriel@x.keinpfusch.netU This user is from outside of this forum
                            uriel@x.keinpfusch.netU This user is from outside of this forum
                            uriel@x.keinpfusch.net
                            wrote last edited by
                            #28

                            @GreenSkyOverMe @EmilyEnough

                            whatever.

                            1 Reply Last reply
                            0
                            • uriel@x.keinpfusch.netU uriel@x.keinpfusch.net

                              @pglpm @EmilyEnough

                              ok, since you aren't able to, let me google for sources:

                              Link Preview Image
                              We did the math on AI’s energy footprint. Here’s the story you haven’t heard.

                              The emissions from individual AI text, image, and video queries seem small—until you add up what the industry isn’t tracking and consider where it’s heading next.

                              favicon

                              MIT Technology Review (www.technologyreview.com)

                              MIT says , 4.4%.

                              Arxiv is so full of shit, I don't even care. WARNING: next time you ask me to google something for you, since you are too stupid for , you must pay me.

                              Link Preview Image
                              pglpm@c.imP This user is from outside of this forum
                              pglpm@c.imP This user is from outside of this forum
                              pglpm@c.im
                              wrote last edited by
                              #29

                              @uriel @EmilyEnough

                              So:
                              - you make claims without supporting evidence,
                              - you simply dismiss as "full of shit" any evidence that's inconvenient to you,
                              - you just call others "stupid".

                              I don't know if you think you're smart, but with these traits the other people see very clearly that you aren't different from a flat-earther, and will treat your claims accordingly. Guess who's the one "full of shit".

                              Bye bye Mr Flat-Earth.

                              1 Reply Last reply
                              0
                              • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

                                My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                metin@graphics.socialM This user is from outside of this forum
                                metin@graphics.socialM This user is from outside of this forum
                                metin@graphics.social
                                wrote last edited by
                                #30

                                @EmilyEnough Well said. This could never have been LLM-generated. 🙂👍

                                1 Reply Last reply
                                0
                                • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

                                  My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                  LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                  In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                  But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                  If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                  greenpete@sunny.gardenG This user is from outside of this forum
                                  greenpete@sunny.gardenG This user is from outside of this forum
                                  greenpete@sunny.garden
                                  wrote last edited by
                                  #31

                                  @EmilyEnough On a slight side note, have you seen this...

                                  Link Preview Image
                                  Athena (@Climatehistories@mastodon.social)

                                  Attached: 1 video Sam Altman: “We see a future where intelligence is a utility, like electricity or water, and people buy it from us on a meter.” Meaning: We stole all your knowledge , writing, and art, and now we’re gonna put a meter on it and sell it back to you. You’re welcome.”

                                  favicon

                                  Mastodon (mastodon.social)

                                  1 Reply Last reply
                                  0
                                  • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

                                    My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                    LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                    In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                    But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                    If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                    chase@chaos.socialC This user is from outside of this forum
                                    chase@chaos.socialC This user is from outside of this forum
                                    chase@chaos.social
                                    wrote last edited by
                                    #32

                                    @EmilyEnough this is a very justified rant

                                    But the thought of computers being too autistic so people had to turn them neurotypical by adding llms is just so funny

                                    1 Reply Last reply
                                    0
                                    • ? Guest

                                      @EmilyEnough Wow, I have thought a lot about how coding LLMs are antithetical to my own OCD tendencies that want everything to be built and formatted in a very specific way (i.e. the right way), but had not considered how terrible the interface would be for folks who prefer not to have to process information conversationally.

                                      I would love to read an entire book or series of articles about how LLMs as an interface enforce neurotypical modes of communication on neurodiverse people.

                                      gourd@indiepocalypse.socialG This user is from outside of this forum
                                      gourd@indiepocalypse.socialG This user is from outside of this forum
                                      gourd@indiepocalypse.social
                                      wrote last edited by
                                      #33

                                      @mikemccaffrey @EmilyEnough The "you can write natural language queries" idea has always gotten a response from me of "why the fuck would I want to do that?" Standard search engine queries and stuff are so much easier.

                                      1 Reply Last reply
                                      0
                                      • emilyenough@hachyderm.ioE emilyenough@hachyderm.io

                                        My biggest problem with the concept of LLMs, even if they weren’t a giant plagiarism laundering machine and disaster for the environment, is that they introduce so much unpredictability into computing. I became a professional computer toucher because they do exactly what you tell them to. Not always what you wanted, but exactly what you asked for.

                                        LLMs turn that upside down. They turn a very autistic do-what-you-say, say-what-you-mean commmunication style with the machine into a neurotypical conversation talking around the issue, but never directly addressing the substance of problem.

                                        In any conversation I have with a person, I’m modeling their understanding of the topic at hand, trying to tailor my communication style to their needs. The same applies to programming languages and frameworks. If you work with a language the way its author intended it goes a lot easier.

                                        But LLMs don’t have an understanding of the conversation. There is no intent. It’s just a mostly-likely-next-word generator on steroids. You’re trying to give directions to a lossily compressed copy of the entire works of human writing. There is no mind to model, and no predictability to the output.

                                        If I wanted to spend my time communicating in a superficial, neurotypical style my autistic ass certainly wouldn’t have gone into computering. LLMs are the final act of the finance bros and capitalists wrestling modern technology away from the technically literate proletariat who built it.

                                        rupert@mastodon.nzR This user is from outside of this forum
                                        rupert@mastodon.nzR This user is from outside of this forum
                                        rupert@mastodon.nz
                                        wrote last edited by
                                        #34

                                        @EmilyEnough Had an interesting chat with the senior director at my office recently. He pointed out that as far as he can see, he already uses natural language to explain what he wants from software. This is just faster.
                                        It was a perspective I hadn't considered before, but the more I think about it the more I think it's deeply insulting.

                                        emilyenough@hachyderm.ioE david_chisnall@infosec.exchangeD 2 Replies Last reply
                                        0
                                        • rupert@mastodon.nzR rupert@mastodon.nz

                                          @EmilyEnough Had an interesting chat with the senior director at my office recently. He pointed out that as far as he can see, he already uses natural language to explain what he wants from software. This is just faster.
                                          It was a perspective I hadn't considered before, but the more I think about it the more I think it's deeply insulting.

                                          emilyenough@hachyderm.ioE This user is from outside of this forum
                                          emilyenough@hachyderm.ioE This user is from outside of this forum
                                          emilyenough@hachyderm.io
                                          wrote last edited by
                                          #35

                                          @rupert he is telling you flat out that he plans on replacing the expensive translation layer (you) asap. By and large that’s how the entire capital class sees this technology, as a way to eliminate expensive human labor without doing any actual work themselves.

                                          rupert@mastodon.nzR 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups