Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

Scheduled Pinned Locked Moved Uncategorized
15 Posts 12 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • cammerman@mstdn.socialC cammerman@mstdn.social

    Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

    This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

    pmonks@sfba.socialP This user is from outside of this forum
    pmonks@sfba.socialP This user is from outside of this forum
    pmonks@sfba.social
    wrote last edited by
    #4

    @cammerman @mhoye I don’t suppose I could impose upon you to give mine a quick review? https://github.com/pmonks/wreck/blob/dev/AGENTS.md

    not2b@sfba.socialN 1 Reply Last reply
    0
    • cammerman@mstdn.socialC cammerman@mstdn.social

      Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

      Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

      pseudonym@mastodon.onlineP This user is from outside of this forum
      pseudonym@mastodon.onlineP This user is from outside of this forum
      pseudonym@mastodon.online
      wrote last edited by
      #5

      @cammerman

      This is an all too common failure mode (of humans). The system is working as designed, being a next most probable token generator.

      What I keep encountering is humans failing to grasp that the LLM has no world model, and no sense whatsoever of "meaning" or "truth" of anything, ever.

      Because a lot of world model, truth-based, reasoning is implicitly encoded in language, frequently true things are probable next tokens.

      This makes humans think "understanding" happens. It doesn't.

      nielsa@mas.toN 1 Reply Last reply
      0
      • cammerman@mstdn.socialC cammerman@mstdn.social

        Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

        Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

        cammerman@mstdn.socialC This user is from outside of this forum
        cammerman@mstdn.socialC This user is from outside of this forum
        cammerman@mstdn.social
        wrote last edited by
        #6

        No matter how many prompt contexts you stack up and fire off in parallel, the machine cannot find truth, cannot do math, cannot know things, or reason.

        It's Massively Multiplayer Online Autocomplete.

        The fact that the capital and executive class thinks this is sufficient to replace most of the world's knowledge workers tells you all you need to know about how we should be dealing with them, and all of this.

        lnklnx@social.lnklnx.comL 1 Reply Last reply
        1
        0
        • pseudonym@mastodon.onlineP pseudonym@mastodon.online

          @cammerman

          This is an all too common failure mode (of humans). The system is working as designed, being a next most probable token generator.

          What I keep encountering is humans failing to grasp that the LLM has no world model, and no sense whatsoever of "meaning" or "truth" of anything, ever.

          Because a lot of world model, truth-based, reasoning is implicitly encoded in language, frequently true things are probable next tokens.

          This makes humans think "understanding" happens. It doesn't.

          nielsa@mas.toN This user is from outside of this forum
          nielsa@mas.toN This user is from outside of this forum
          nielsa@mas.to
          wrote last edited by
          #7

          @pseudonym @cammerman If you ask the magic 8-ball about frequently truthfully described or discussed things, or things of similar structure with a clear mapping, it's more likely to produce a correct answer, but you have no idea what's frequently described and whether it's just randomly wrong. Oh, and if it is wrong, it is optimized to make the wrong answer look right in context. Good luck.

          1 Reply Last reply
          0
          • pmonks@sfba.socialP pmonks@sfba.social

            @cammerman @mhoye I don’t suppose I could impose upon you to give mine a quick review? https://github.com/pmonks/wreck/blob/dev/AGENTS.md

            not2b@sfba.socialN This user is from outside of this forum
            not2b@sfba.socialN This user is from outside of this forum
            not2b@sfba.social
            wrote last edited by
            #8

            @pmonks @cammerman @mhoye I'm curious to see some of the limericks that are generated by these instructions.

            1 Reply Last reply
            0
            • cammerman@mstdn.socialC This user is from outside of this forum
              cammerman@mstdn.socialC This user is from outside of this forum
              cammerman@mstdn.social
              wrote last edited by
              #9

              @pmonks @not2b @mhoye

              😘👌 No notes.

              1 Reply Last reply
              0
              • cammerman@mstdn.socialC cammerman@mstdn.social

                Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

                Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

                walrus@toot.walesW This user is from outside of this forum
                walrus@toot.walesW This user is from outside of this forum
                walrus@toot.wales
                wrote last edited by
                #10

                @cammerman

                They don't KNOW anything, full stop.

                They're just autocorrect on steroids.

                1 Reply Last reply
                0
                • cammerman@mstdn.socialC cammerman@mstdn.social

                  Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

                  Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

                  individeweal@mas.toI This user is from outside of this forum
                  individeweal@mas.toI This user is from outside of this forum
                  individeweal@mas.to
                  wrote last edited by
                  #11

                  @cammerman Possibly more important, it doesn't know what "don't" means. I'm not even making a philosophical point about what it is to "know" something, even if we only care about output, it doesn't act in a way that corresponds to following an instruction not to do anything. Don't mention goblins: no effect on how often goblins get mentioned if the training was weighted towards mentioning creatures.

                  rndanger@infosec.exchangeR 1 Reply Last reply
                  0
                  • cammerman@mstdn.socialC cammerman@mstdn.social

                    Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

                    This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

                    gildilinie@beige.partyG This user is from outside of this forum
                    gildilinie@beige.partyG This user is from outside of this forum
                    gildilinie@beige.party
                    wrote last edited by
                    #12

                    @cammerman https://beige.party/@gildilinie/116500174248274955

                    1 Reply Last reply
                    0
                    • individeweal@mas.toI individeweal@mas.to

                      @cammerman Possibly more important, it doesn't know what "don't" means. I'm not even making a philosophical point about what it is to "know" something, even if we only care about output, it doesn't act in a way that corresponds to following an instruction not to do anything. Don't mention goblins: no effect on how often goblins get mentioned if the training was weighted towards mentioning creatures.

                      rndanger@infosec.exchangeR This user is from outside of this forum
                      rndanger@infosec.exchangeR This user is from outside of this forum
                      rndanger@infosec.exchange
                      wrote last edited by
                      #13

                      @individeweal @cammerman
                      I told ChatGPT "don't apologize" so it apologized for apologizing.

                      1 Reply Last reply
                      0
                      • cammerman@mstdn.socialC cammerman@mstdn.social

                        Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

                        This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

                        petersommerlad@mastodon.socialP This user is from outside of this forum
                        petersommerlad@mastodon.socialP This user is from outside of this forum
                        petersommerlad@mastodon.social
                        wrote last edited by
                        #14

                        Retterin der Autoindustrie (Chinas)

                        1 Reply Last reply
                        0
                        • cammerman@mstdn.socialC cammerman@mstdn.social

                          No matter how many prompt contexts you stack up and fire off in parallel, the machine cannot find truth, cannot do math, cannot know things, or reason.

                          It's Massively Multiplayer Online Autocomplete.

                          The fact that the capital and executive class thinks this is sufficient to replace most of the world's knowledge workers tells you all you need to know about how we should be dealing with them, and all of this.

                          lnklnx@social.lnklnx.comL This user is from outside of this forum
                          lnklnx@social.lnklnx.comL This user is from outside of this forum
                          lnklnx@social.lnklnx.com
                          wrote last edited by
                          #15

                          @cammerman "Massively Multiplayer Online Autocomplete", lol

                          I am officially calling it this from now on!

                          1 Reply Last reply
                          0
                          • R relay@relay.mycrowd.ca shared this topic
                          Reply
                          • Reply as topic
                          Log in to reply
                          • Oldest to Newest
                          • Newest to Oldest
                          • Most Votes


                          • Login

                          • Login or register to search.
                          • First post
                            Last post
                          0
                          • Categories
                          • Recent
                          • Tags
                          • Popular
                          • World
                          • Users
                          • Groups