Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

Scheduled Pinned Locked Moved Uncategorized
15 Posts 12 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • cammerman@mstdn.socialC This user is from outside of this forum
    cammerman@mstdn.socialC This user is from outside of this forum
    cammerman@mstdn.social
    wrote last edited by
    #1

    Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

    This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

    cammerman@mstdn.socialC ardvaark@mastodon.worldA pmonks@sfba.socialP gildilinie@beige.partyG petersommerlad@mastodon.socialP 5 Replies Last reply
    0
    • cammerman@mstdn.socialC cammerman@mstdn.social

      Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

      This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

      cammerman@mstdn.socialC This user is from outside of this forum
      cammerman@mstdn.socialC This user is from outside of this forum
      cammerman@mstdn.social
      wrote last edited by
      #2

      Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

      Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

      pseudonym@mastodon.onlineP cammerman@mstdn.socialC walrus@toot.walesW individeweal@mas.toI 4 Replies Last reply
      0
      • cammerman@mstdn.socialC cammerman@mstdn.social

        Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

        This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

        ardvaark@mastodon.worldA This user is from outside of this forum
        ardvaark@mastodon.worldA This user is from outside of this forum
        ardvaark@mastodon.world
        wrote last edited by
        #3

        @cammerman I'm at best 50/50 on what I need to build before I start making changes. ๐Ÿ˜…

        1 Reply Last reply
        0
        • cammerman@mstdn.socialC cammerman@mstdn.social

          Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

          This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

          pmonks@sfba.socialP This user is from outside of this forum
          pmonks@sfba.socialP This user is from outside of this forum
          pmonks@sfba.social
          wrote last edited by
          #4

          @cammerman @mhoye I donโ€™t suppose I could impose upon you to give mine a quick review? https://github.com/pmonks/wreck/blob/dev/AGENTS.md

          not2b@sfba.socialN 1 Reply Last reply
          0
          • cammerman@mstdn.socialC cammerman@mstdn.social

            Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

            Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

            pseudonym@mastodon.onlineP This user is from outside of this forum
            pseudonym@mastodon.onlineP This user is from outside of this forum
            pseudonym@mastodon.online
            wrote last edited by
            #5

            @cammerman

            This is an all too common failure mode (of humans). The system is working as designed, being a next most probable token generator.

            What I keep encountering is humans failing to grasp that the LLM has no world model, and no sense whatsoever of "meaning" or "truth" of anything, ever.

            Because a lot of world model, truth-based, reasoning is implicitly encoded in language, frequently true things are probable next tokens.

            This makes humans think "understanding" happens. It doesn't.

            nielsa@mas.toN 1 Reply Last reply
            0
            • cammerman@mstdn.socialC cammerman@mstdn.social

              Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

              Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

              cammerman@mstdn.socialC This user is from outside of this forum
              cammerman@mstdn.socialC This user is from outside of this forum
              cammerman@mstdn.social
              wrote last edited by
              #6

              No matter how many prompt contexts you stack up and fire off in parallel, the machine cannot find truth, cannot do math, cannot know things, or reason.

              It's Massively Multiplayer Online Autocomplete.

              The fact that the capital and executive class thinks this is sufficient to replace most of the world's knowledge workers tells you all you need to know about how we should be dealing with them, and all of this.

              lnklnx@social.lnklnx.comL 1 Reply Last reply
              1
              0
              • pseudonym@mastodon.onlineP pseudonym@mastodon.online

                @cammerman

                This is an all too common failure mode (of humans). The system is working as designed, being a next most probable token generator.

                What I keep encountering is humans failing to grasp that the LLM has no world model, and no sense whatsoever of "meaning" or "truth" of anything, ever.

                Because a lot of world model, truth-based, reasoning is implicitly encoded in language, frequently true things are probable next tokens.

                This makes humans think "understanding" happens. It doesn't.

                nielsa@mas.toN This user is from outside of this forum
                nielsa@mas.toN This user is from outside of this forum
                nielsa@mas.to
                wrote last edited by
                #7

                @pseudonym @cammerman If you ask the magic 8-ball about frequently truthfully described or discussed things, or things of similar structure with a clear mapping, it's more likely to produce a correct answer, but you have no idea what's frequently described and whether it's just randomly wrong. Oh, and if it is wrong, it is optimized to make the wrong answer look right in context. Good luck.

                1 Reply Last reply
                0
                • pmonks@sfba.socialP pmonks@sfba.social

                  @cammerman @mhoye I donโ€™t suppose I could impose upon you to give mine a quick review? https://github.com/pmonks/wreck/blob/dev/AGENTS.md

                  not2b@sfba.socialN This user is from outside of this forum
                  not2b@sfba.socialN This user is from outside of this forum
                  not2b@sfba.social
                  wrote last edited by
                  #8

                  @pmonks @cammerman @mhoye I'm curious to see some of the limericks that are generated by these instructions.

                  1 Reply Last reply
                  0
                  • cammerman@mstdn.socialC This user is from outside of this forum
                    cammerman@mstdn.socialC This user is from outside of this forum
                    cammerman@mstdn.social
                    wrote last edited by
                    #9

                    @pmonks @not2b @mhoye

                    ๐Ÿ˜˜๐Ÿ‘Œ No notes.

                    1 Reply Last reply
                    0
                    • cammerman@mstdn.socialC cammerman@mstdn.social

                      Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

                      Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

                      walrus@toot.walesW This user is from outside of this forum
                      walrus@toot.walesW This user is from outside of this forum
                      walrus@toot.wales
                      wrote last edited by
                      #10

                      @cammerman

                      They don't KNOW anything, full stop.

                      They're just autocorrect on steroids.

                      1 Reply Last reply
                      0
                      • cammerman@mstdn.socialC cammerman@mstdn.social

                        Look, if you tell an LLM it needs 95% confidence, it doesn't know what either "95%" nor "confidence" means. It knows people tend to respond to this kind of direction either by saying "I'm not sure enough because..." or "I'm super confident for these reasons." It has no ability to correctly choose which of those templates it will follow.

                        Flip a coin. You'll get a reasonable looking sentence back in one of those styles, with a random assortment of reasons that may or may not be rooted in fact.

                        individeweal@mas.toI This user is from outside of this forum
                        individeweal@mas.toI This user is from outside of this forum
                        individeweal@mas.to
                        wrote last edited by
                        #11

                        @cammerman Possibly more important, it doesn't know what "don't" means. I'm not even making a philosophical point about what it is to "know" something, even if we only care about output, it doesn't act in a way that corresponds to following an instruction not to do anything. Don't mention goblins: no effect on how often goblins get mentioned if the training was weighted towards mentioning creatures.

                        rndanger@infosec.exchangeR 1 Reply Last reply
                        0
                        • cammerman@mstdn.socialC cammerman@mstdn.social

                          Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

                          This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

                          gildilinie@beige.partyG This user is from outside of this forum
                          gildilinie@beige.partyG This user is from outside of this forum
                          gildilinie@beige.party
                          wrote last edited by
                          #12

                          @cammerman https://beige.party/@gildilinie/116500174248274955

                          1 Reply Last reply
                          0
                          • individeweal@mas.toI individeweal@mas.to

                            @cammerman Possibly more important, it doesn't know what "don't" means. I'm not even making a philosophical point about what it is to "know" something, even if we only care about output, it doesn't act in a way that corresponds to following an instruction not to do anything. Don't mention goblins: no effect on how often goblins get mentioned if the training was weighted towards mentioning creatures.

                            rndanger@infosec.exchangeR This user is from outside of this forum
                            rndanger@infosec.exchangeR This user is from outside of this forum
                            rndanger@infosec.exchange
                            wrote last edited by
                            #13

                            @individeweal @cammerman
                            I told ChatGPT "don't apologize" so it apologized for apologizing.

                            1 Reply Last reply
                            0
                            • cammerman@mstdn.socialC cammerman@mstdn.social

                              Reading a claude.md and seeing written there many directives like this: "Don't make changes until you have 95% confidence in what you need to build."

                              This reveals such a profound misunderstanding of how this technology works that I'm speechless. And this is literally what people are trying to build fully-automated "software factories" from.

                              petersommerlad@mastodon.socialP This user is from outside of this forum
                              petersommerlad@mastodon.socialP This user is from outside of this forum
                              petersommerlad@mastodon.social
                              wrote last edited by
                              #14

                              Retterin der Autoindustrie (Chinas)

                              1 Reply Last reply
                              0
                              • cammerman@mstdn.socialC cammerman@mstdn.social

                                No matter how many prompt contexts you stack up and fire off in parallel, the machine cannot find truth, cannot do math, cannot know things, or reason.

                                It's Massively Multiplayer Online Autocomplete.

                                The fact that the capital and executive class thinks this is sufficient to replace most of the world's knowledge workers tells you all you need to know about how we should be dealing with them, and all of this.

                                lnklnx@social.lnklnx.comL This user is from outside of this forum
                                lnklnx@social.lnklnx.comL This user is from outside of this forum
                                lnklnx@social.lnklnx.com
                                wrote last edited by
                                #15

                                @cammerman "Massively Multiplayer Online Autocomplete", lol

                                I am officially calling it this from now on!

                                1 Reply Last reply
                                0
                                • R relay@relay.mycrowd.ca shared this topic
                                Reply
                                • Reply as topic
                                Log in to reply
                                • Oldest to Newest
                                • Newest to Oldest
                                • Most Votes


                                • Login

                                • Login or register to search.
                                • First post
                                  Last post
                                0
                                • Categories
                                • Recent
                                • Tags
                                • Popular
                                • World
                                • Users
                                • Groups