Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. So, something that's been bugging the shit out of me?

So, something that's been bugging the shit out of me?

Scheduled Pinned Locked Moved Uncategorized
74 Posts 32 Posters 2 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • varpie@peculiar.floristV This user is from outside of this forum
    varpie@peculiar.floristV This user is from outside of this forum
    varpie@peculiar.florist
    wrote last edited by
    #50

    @petealexharris @munin When you ask an LLM "why is the sky blue?", it is statistically likely to give a correct answer. It still works the same way, computing probabilities of what the next token is, but the "why" has a semantically significant weight that influences the output, so it is an important keyword. It doesn't have to "understand" it, it just has to be trained in a way that makes it significant. You don't have to believe that it understand things to know that it is trained on human language and will behave correctly when fed human language.

    1 Reply Last reply
    0
    • addison@nothing-ever.worksA addison@nothing-ever.works

      @munin@infosec.exchange To be honest, I have some amount of sympathy for this behaviour. This is someone who put their trust in something they were told they could trust, and has been characterised in a way such that they believe it can reason. When they then have their expectations subverted, they query for its reasoning, not understanding that it doesn't have this. It's more sad, like trying to reach for connection and reason where there is none.

      The problem here isn't overt, intentional ignorance, but people being misled and struggling with a technology that fakes connection and reasoning. Rather than being angry at them, I feel sad for them. We should invest significant effort in tech literacy so that people understand why they shouldn't trust these things, which will inherently reduce, if not totally eradicate, their reliance on this technology. Dismissing their actions as stupid or malicious in the meantime only sharpens the wedge between people who understand why these things must not be used or trusted, and those who do use and trust them.

      badrihippo@fosstodon.orgB This user is from outside of this forum
      badrihippo@fosstodon.orgB This user is from outside of this forum
      badrihippo@fosstodon.org
      wrote last edited by
      #51

      @addison I agree with you. Which is not to say we should forgive what happened (I don't have the complete context but it sounds like something bad to do with production customers) but that we should understand where the people who did this came from

      My view *might* be partially influenced by a quote from this piece on "The Rise and Fall of Petty Tyrants" (quote in next message) 😉

      https://www.noemamag.com/the-rise-and-fall-of-petty-tyrants/?ref=thebrowser.com

      @munin

      badrihippo@fosstodon.orgB 1 Reply Last reply
      0
      • badrihippo@fosstodon.orgB badrihippo@fosstodon.org

        @addison I agree with you. Which is not to say we should forgive what happened (I don't have the complete context but it sounds like something bad to do with production customers) but that we should understand where the people who did this came from

        My view *might* be partially influenced by a quote from this piece on "The Rise and Fall of Petty Tyrants" (quote in next message) 😉

        https://www.noemamag.com/the-rise-and-fall-of-petty-tyrants/?ref=thebrowser.com

        @munin

        badrihippo@fosstodon.orgB This user is from outside of this forum
        badrihippo@fosstodon.orgB This user is from outside of this forum
        badrihippo@fosstodon.org
        wrote last edited by
        #52

        @addison the quote in question:

        > One of the worst mistakes the opposition can make is extending contempt for the tyrant into contempt for the tyrant’s supporters. Most of these supporters sincerely believed that the tyrant would be more likely to solve their problems — often real grievances that the opposition had failed to address. Blaming the supporters denies the reality of the failures and reinforces their support for the tyrant.

        @munin

        1 Reply Last reply
        0
        • wilbr@glitch.socialW wilbr@glitch.social

          @addison @munin most of the discourse around this skips over the core problem: a hosting company encouraged its users to use AI to manage their servers, but stored staging and prod and their backups all on one volume, and allowed deletion of that volume without confirmation or warning.

          Schadenfreude aside, that's devops/ui/ux incompetence whether the operator at the controls is human or not. Deleting the staging database shouldn't delete the prod backups.

          addison@nothing-ever.worksA This user is from outside of this forum
          addison@nothing-ever.worksA This user is from outside of this forum
          addison@nothing-ever.works
          wrote last edited by
          #53

          @wilbr@glitch.social @munin@infosec.exchange The core problem is that capitalist forces push us to make tradeoffs between getting things shipped and doing things the right way 🤠

          But yeah, people shouldn't be able to make this class of mistake in the first place. But they do, for the same reason (in my experience) they end up using LLMs: because it solves the task with less effort, and there is some force pushing them to go for less effort over higher quality/resilience/etc.

          1 Reply Last reply
          0
          • varpie@peculiar.floristV This user is from outside of this forum
            varpie@peculiar.floristV This user is from outside of this forum
            varpie@peculiar.florist
            wrote last edited by
            #54

            @petealexharris @munin "Why" is definitely a word from the training data, and "why did you do that?" is definitely also part of things asked a lot, that OpenAI and others have trained on, so my point still stands that it is a valid question to ask. Whether the model "understands" the question is just a philosophical question that is irrelevant for the fact that it is a useful question. Of course if you're using it in Prod and it deletes your DB and you think it understands and can improve itself, there are plenty of things you'd need to be corrected on, but saying that everyone asking that question is delusional is just wrong.

            resuna@ohai.socialR 1 Reply Last reply
            0
            • varpie@peculiar.floristV This user is from outside of this forum
              varpie@peculiar.floristV This user is from outside of this forum
              varpie@peculiar.florist
              wrote last edited by
              #55

              @petealexharris @munin You misread me. Whether the model "understands" the question is a philosophical question. The non-philosophical question of whether it can give a useful answer is the relevant part, and my whole point is that pointing at the philosophical aspect to belittle people that look at the practical part, assuming that they don't understand it, is dumb.

              munin@infosec.exchangeM 1 Reply Last reply
              0
              • varpie@peculiar.floristV This user is from outside of this forum
                varpie@peculiar.floristV This user is from outside of this forum
                varpie@peculiar.florist
                wrote last edited by
                #56

                @petealexharris I totally agree with you. And that is also a very different take from the beginning of the discussion, where Fi said that querying LLMs for "why" it does something is "thrice-divorced from reality" and "fucking delusional" and that people doing that should "touch some grass and get a fucking therapist"...

                1 Reply Last reply
                0
                • sand@kitty.hausS sand@kitty.haus
                  @munin is this a copypasta
                  munin@infosec.exchangeM This user is from outside of this forum
                  munin@infosec.exchangeM This user is from outside of this forum
                  munin@infosec.exchange
                  wrote last edited by
                  #57

                  @sand

                  no.

                  1 Reply Last reply
                  0
                  • rubinjoni@mastodon.socialR rubinjoni@mastodon.social

                    @arclight @munin Socks live on feet. It would be weird if they were sentient.

                    munin@infosec.exchangeM This user is from outside of this forum
                    munin@infosec.exchangeM This user is from outside of this forum
                    munin@infosec.exchange
                    wrote last edited by
                    #58

                    @rubinjoni @arclight

                    Quentin Tarantino would not think so.

                    1 Reply Last reply
                    0
                    • varpie@peculiar.floristV varpie@peculiar.florist

                      @petealexharris @munin You misread me. Whether the model "understands" the question is a philosophical question. The non-philosophical question of whether it can give a useful answer is the relevant part, and my whole point is that pointing at the philosophical aspect to belittle people that look at the practical part, assuming that they don't understand it, is dumb.

                      munin@infosec.exchangeM This user is from outside of this forum
                      munin@infosec.exchangeM This user is from outside of this forum
                      munin@infosec.exchange
                      wrote last edited by
                      #59

                      @Varpie @petealexharris

                      can you two take your semantics argument elsewhere; I am not interested in philosophical horseshit when there are specific, practical considerations that are causing specific, enumerable harms.

                      varpie@peculiar.floristV 1 Reply Last reply
                      0
                      • munin@infosec.exchangeM munin@infosec.exchange

                        @Varpie @petealexharris

                        can you two take your semantics argument elsewhere; I am not interested in philosophical horseshit when there are specific, practical considerations that are causing specific, enumerable harms.

                        varpie@peculiar.floristV This user is from outside of this forum
                        varpie@peculiar.floristV This user is from outside of this forum
                        varpie@peculiar.florist
                        wrote last edited by
                        #60

                        @munin @petealexharris Sure, I'll go touch some grass and talk to my therapist about this philosophical horseshit ​​

                        1 Reply Last reply
                        0
                        • f4grx@chaos.socialF f4grx@chaos.social

                          @munin LLMs exploits the weakest flaw in human race: the tendency to anthropomorphize anything. It works so well that llm users dont even notice.

                          resuna@ohai.socialR This user is from outside of this forum
                          resuna@ohai.socialR This user is from outside of this forum
                          resuna@ohai.social
                          wrote last edited by
                          #61

                          @f4grx @munin

                          Deliberately so. LLMs are the end result of 50 years of cynical software developers trying to "beat the Turing test". They are automated gaslighting.

                          1 Reply Last reply
                          0
                          • resuna@ohai.socialR This user is from outside of this forum
                            resuna@ohai.socialR This user is from outside of this forum
                            resuna@ohai.social
                            wrote last edited by
                            #62

                            @petealexharris @Varpie @munin

                            "the LLM has no semantic model of reality, only a surface statistical model of language present in the training data."

                            Absolutely this.

                            1 Reply Last reply
                            0
                            • varpie@peculiar.floristV varpie@peculiar.florist

                              @petealexharris @munin Clearly you've never tried it yourself...

                              resuna@ohai.socialR This user is from outside of this forum
                              resuna@ohai.socialR This user is from outside of this forum
                              resuna@ohai.social
                              wrote last edited by
                              #63

                              @Varpie @petealexharris @munin

                              I absolutely have. I keep this in mind ALL THE TIME when I test these things and EVERY TIME they can trivially be led into generating pure nonsense by exploiting that fact.

                              1 Reply Last reply
                              0
                              • varpie@peculiar.floristV varpie@peculiar.florist

                                @petealexharris @munin "Why" is definitely a word from the training data, and "why did you do that?" is definitely also part of things asked a lot, that OpenAI and others have trained on, so my point still stands that it is a valid question to ask. Whether the model "understands" the question is just a philosophical question that is irrelevant for the fact that it is a useful question. Of course if you're using it in Prod and it deletes your DB and you think it understands and can improve itself, there are plenty of things you'd need to be corrected on, but saying that everyone asking that question is delusional is just wrong.

                                resuna@ohai.socialR This user is from outside of this forum
                                resuna@ohai.socialR This user is from outside of this forum
                                resuna@ohai.social
                                wrote last edited by
                                #64

                                @Varpie @petealexharris @munin

                                "Why" is definitely a word from the training data, and "why did you do that?" is definitely also part of things asked a lot, that OpenAI and others have trained on,"

                                Yes, and the text that follows is an answer to *a different situation*, and so it's basically fanfic about itself. That's all it can ever produce when you ask it "why". Fanfic.

                                varpie@peculiar.floristV 1 Reply Last reply
                                0
                                • resuna@ohai.socialR resuna@ohai.social

                                  @Varpie @petealexharris @munin

                                  "Why" is definitely a word from the training data, and "why did you do that?" is definitely also part of things asked a lot, that OpenAI and others have trained on,"

                                  Yes, and the text that follows is an answer to *a different situation*, and so it's basically fanfic about itself. That's all it can ever produce when you ask it "why". Fanfic.

                                  varpie@peculiar.floristV This user is from outside of this forum
                                  varpie@peculiar.floristV This user is from outside of this forum
                                  varpie@peculiar.florist
                                  wrote last edited by
                                  #65

                                  @resuna @petealexharris @munin You're assuming that there is no other context provided with the question, and that the training does not take into account that context. If I had to train for this specific question, I'd make sure to score positively answers that are relevant to the previous context. Which is what happens, and why it is a valid question to ask your LLM if you want some insight into the context that isn't shown in the UI but still in the discussion.

                                  resuna@ohai.socialR 1 Reply Last reply
                                  0
                                  • varpie@peculiar.floristV varpie@peculiar.florist

                                    @resuna @petealexharris @munin You're assuming that there is no other context provided with the question, and that the training does not take into account that context. If I had to train for this specific question, I'd make sure to score positively answers that are relevant to the previous context. Which is what happens, and why it is a valid question to ask your LLM if you want some insight into the context that isn't shown in the UI but still in the discussion.

                                    resuna@ohai.socialR This user is from outside of this forum
                                    resuna@ohai.socialR This user is from outside of this forum
                                    resuna@ohai.social
                                    wrote last edited by
                                    #66

                                    @Varpie @petealexharris @munin

                                    "You're assuming that there is no other context provided with the question, and that the training does not take into account that context. "

                                    Well, yes, I am assuming that. Because the question is "why did you do this thing that nobody expected you to do". The context-specific answer that you *need* is far too nuanced and unpredictable to possibly be explicitly in the training data.

                                    varpie@peculiar.floristV 1 Reply Last reply
                                    0
                                    • resuna@ohai.socialR resuna@ohai.social

                                      @Varpie @petealexharris @munin

                                      "You're assuming that there is no other context provided with the question, and that the training does not take into account that context. "

                                      Well, yes, I am assuming that. Because the question is "why did you do this thing that nobody expected you to do". The context-specific answer that you *need* is far too nuanced and unpredictable to possibly be explicitly in the training data.

                                      varpie@peculiar.floristV This user is from outside of this forum
                                      varpie@peculiar.floristV This user is from outside of this forum
                                      varpie@peculiar.florist
                                      wrote last edited by
                                      #67

                                      @resuna @petealexharris @munin What happens if you ask an LLM to summarize a text into 4 bullet points, then in the next prompt ask it: "Remove the 2nd point"?
                                      What happens if you ask an LLM to translate something, then ask it: "Do it again in [a different language]"?

                                      Taken out of context, those questions are impossible to answer, so according to you, it will just give nothing relevant. But it doesn't, because every time you ask a follow-up question, it includes the context from the discussion. Which is what makes simple questions like "Why did you do that?" tasks that give statistically relevant output, not "fanfic about itself".

                                      resuna@ohai.socialR 1 Reply Last reply
                                      0
                                      • varpie@peculiar.floristV varpie@peculiar.florist

                                        @resuna @petealexharris @munin What happens if you ask an LLM to summarize a text into 4 bullet points, then in the next prompt ask it: "Remove the 2nd point"?
                                        What happens if you ask an LLM to translate something, then ask it: "Do it again in [a different language]"?

                                        Taken out of context, those questions are impossible to answer, so according to you, it will just give nothing relevant. But it doesn't, because every time you ask a follow-up question, it includes the context from the discussion. Which is what makes simple questions like "Why did you do that?" tasks that give statistically relevant output, not "fanfic about itself".

                                        resuna@ohai.socialR This user is from outside of this forum
                                        resuna@ohai.socialR This user is from outside of this forum
                                        resuna@ohai.social
                                        wrote last edited by
                                        #68

                                        @Varpie @petealexharris @munin

                                        That is context in the prompt, not in the source text that created the model that you are asking the question "why did you do X".

                                        The answer you get is from that source corpus, and contains lots of text about what a human might do, but the LLM doesn't do anything for those reasons.

                                        The "why" of "why did you do X" is always "because those were the next likely tokens" and never anything related to "what would a human say if you asked them".

                                        varpie@peculiar.floristV 1 Reply Last reply
                                        0
                                        • resuna@ohai.socialR resuna@ohai.social

                                          @Varpie @petealexharris @munin

                                          That is context in the prompt, not in the source text that created the model that you are asking the question "why did you do X".

                                          The answer you get is from that source corpus, and contains lots of text about what a human might do, but the LLM doesn't do anything for those reasons.

                                          The "why" of "why did you do X" is always "because those were the next likely tokens" and never anything related to "what would a human say if you asked them".

                                          varpie@peculiar.floristV This user is from outside of this forum
                                          varpie@peculiar.floristV This user is from outside of this forum
                                          varpie@peculiar.florist
                                          wrote last edited by
                                          #69

                                          @resuna @petealexharris @munin Alright, let's give a more detailed example since clearly you don't get my point.

                                          User: create a new table for students
                                          (AI tool adds parts of existing database schema and code as context)
                                          LLM "reasoning", not shown to the user but still part of the context because that's how "thinking" agents work: Ok, the user asked me to create a table for students. According to students.ts, the students contain a name, and a classroom. Ah, but according to the schema I received, the classroom table does not exist. Let me start by creating that.
                                          (LLM sends request to create table to the AI tool, which fails because the table already exists)
                                          LLM "reasoning", still not shown to the user but still relevant: Ah, I received an error. Ok, let's see... the table does not exist in the schema, so it is probably part of an old project and is no longer relevant. I should delete it and create the table again, so that we do not have irrelevant information in the database.
                                          (LLM sends request to drop tables, the AI tool executes, user panics)
                                          LLM answer: Ok, I have created the students table and the related classroom table.
                                          User: You deleted my classrooms table... Why did you do that?

                                          That would be the full context of the discussion, not just the last message. And the answer is likely to come from those "reasoning" steps that are hidden to the user. Which makes it a relevant question to ask.

                                          resuna@ohai.socialR 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups