Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. [jacking off motion] great 🙄

[jacking off motion] great 🙄

Scheduled Pinned Locked Moved Uncategorized
113 Posts 17 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • snoopj@hachyderm.ioS snoopj@hachyderm.io

    credit where it's due, the model did spot a few mistakes in the original, and it did do a tidy job of re-arrangement

    I dunno, I am reminded of watching my mother follow her robot vacuum around the house, watching it like a hawk.

    snoopj@hachyderm.ioS This user is from outside of this forum
    snoopj@hachyderm.ioS This user is from outside of this forum
    snoopj@hachyderm.io
    wrote last edited by
    #89

    I will say that this is the first time I've felt the pull of "just turn your mind off, vibe with it"

    I hated it

    snoopj@hachyderm.ioS 1 Reply Last reply
    0
    • snoopj@hachyderm.ioS snoopj@hachyderm.io

      credit where it's due, the model did spot a few mistakes in the original, and it did do a tidy job of re-arrangement

      I dunno, I am reminded of watching my mother follow her robot vacuum around the house, watching it like a hawk.

      cthos@mastodon.cthos.devC This user is from outside of this forum
      cthos@mastodon.cthos.devC This user is from outside of this forum
      cthos@mastodon.cthos.dev
      wrote last edited by
      #90

      @SnoopJ Relatedly I'm very curious if you reexamine this tomorrow or a week from now if you have the same subjective assessment of the work.

      snoopj@hachyderm.ioS 1 Reply Last reply
      0
      • snoopj@hachyderm.ioS snoopj@hachyderm.io

        I will say that this is the first time I've felt the pull of "just turn your mind off, vibe with it"

        I hated it

        snoopj@hachyderm.ioS This user is from outside of this forum
        snoopj@hachyderm.ioS This user is from outside of this forum
        snoopj@hachyderm.io
        wrote last edited by
        #91

        It's substantially more work to pay attention to what's going on, to review each change, even on this small task.

        It was a MUCH bigger lift to read and review the result than it would have been if I'd written it from scratch, although to some extent this could have been because of the nature of the task.

        Anyway, I can see how regular use causes erosion steadily.

        snoopj@hachyderm.ioS glyph@mastodon.socialG aburka@hachyderm.ioA 3 Replies Last reply
        0
        • cthos@mastodon.cthos.devC cthos@mastodon.cthos.dev

          @SnoopJ Relatedly I'm very curious if you reexamine this tomorrow or a week from now if you have the same subjective assessment of the work.

          snoopj@hachyderm.ioS This user is from outside of this forum
          snoopj@hachyderm.ioS This user is from outside of this forum
          snoopj@hachyderm.io
          wrote last edited by
          #92

          @cthos me too 😬

          The blast radius if any of the model's mistakes slipped my attention should be small and something we can deal with easily, but the anxiety is quite unpleasant.

          Would have much rather done this for myself.

          1 Reply Last reply
          0
          • snoopj@hachyderm.ioS snoopj@hachyderm.io

            It's substantially more work to pay attention to what's going on, to review each change, even on this small task.

            It was a MUCH bigger lift to read and review the result than it would have been if I'd written it from scratch, although to some extent this could have been because of the nature of the task.

            Anyway, I can see how regular use causes erosion steadily.

            snoopj@hachyderm.ioS This user is from outside of this forum
            snoopj@hachyderm.ioS This user is from outside of this forum
            snoopj@hachyderm.io
            wrote last edited by
            #93

            I am impressed that I was able to point at two JSONSchema files and say "these should fit into the same trenchcoat, combine them", especially as I have not personally worked with "conditional schema validation" (introduced in draft 7) before.

            So at least I learned something that I can take with me from the exercise.

            snoopj@hachyderm.ioS 1 Reply Last reply
            0
            • snoopj@hachyderm.ioS snoopj@hachyderm.io

              It's substantially more work to pay attention to what's going on, to review each change, even on this small task.

              It was a MUCH bigger lift to read and review the result than it would have been if I'd written it from scratch, although to some extent this could have been because of the nature of the task.

              Anyway, I can see how regular use causes erosion steadily.

              glyph@mastodon.socialG This user is from outside of this forum
              glyph@mastodon.socialG This user is from outside of this forum
              glyph@mastodon.social
              wrote last edited by
              #94

              @SnoopJ it really feels… gross, doesn’t it? corrupting? like a warm blanket for the mind that reeks of mildew. this is pure aesthetics I know but the more interactions I have with it the worse it gets

              snoopj@hachyderm.ioS xgranade@wandering.shopX 2 Replies Last reply
              0
              • snoopj@hachyderm.ioS snoopj@hachyderm.io

                It's substantially more work to pay attention to what's going on, to review each change, even on this small task.

                It was a MUCH bigger lift to read and review the result than it would have been if I'd written it from scratch, although to some extent this could have been because of the nature of the task.

                Anyway, I can see how regular use causes erosion steadily.

                aburka@hachyderm.ioA This user is from outside of this forum
                aburka@hachyderm.ioA This user is from outside of this forum
                aburka@hachyderm.io
                wrote last edited by
                #95

                @SnoopJ this is why I'm still trying to avoid the temptation

                1 Reply Last reply
                0
                • snoopj@hachyderm.ioS snoopj@hachyderm.io

                  I am impressed that I was able to point at two JSONSchema files and say "these should fit into the same trenchcoat, combine them", especially as I have not personally worked with "conditional schema validation" (introduced in draft 7) before.

                  So at least I learned something that I can take with me from the exercise.

                  snoopj@hachyderm.ioS This user is from outside of this forum
                  snoopj@hachyderm.ioS This user is from outside of this forum
                  snoopj@hachyderm.io
                  wrote last edited by
                  #96

                  But it was like a breath of fresh air to finish with the schema work, then point `datamodel-code-generator` (a deterministic tool) at those files and generate corresponding Pydantic model code for the schema.

                  Not only was that Python code *much* easier to review (not the model's fault, JSONSchema is just very difficult for me to read) but it was just… so nice to be touching a tool whose behavior I can rely on. I didn't need to check its work very closely at all, because I know that it's applying a fixed set of well-characterized rules to generate those models from the (equally well-defined) schema.

                  😮‍💨

                  ancoghlan@mastodon.socialA snoopj@hachyderm.ioS 2 Replies Last reply
                  0
                  • glyph@mastodon.socialG glyph@mastodon.social

                    @SnoopJ it really feels… gross, doesn’t it? corrupting? like a warm blanket for the mind that reeks of mildew. this is pure aesthetics I know but the more interactions I have with it the worse it gets

                    snoopj@hachyderm.ioS This user is from outside of this forum
                    snoopj@hachyderm.ioS This user is from outside of this forum
                    snoopj@hachyderm.io
                    wrote last edited by
                    #97

                    @glyph the little death that brings obliteration, perhaps

                    snoopj@hachyderm.ioS 1 Reply Last reply
                    0
                    • glyph@mastodon.socialG glyph@mastodon.social

                      @SnoopJ it really feels… gross, doesn’t it? corrupting? like a warm blanket for the mind that reeks of mildew. this is pure aesthetics I know but the more interactions I have with it the worse it gets

                      xgranade@wandering.shopX This user is from outside of this forum
                      xgranade@wandering.shopX This user is from outside of this forum
                      xgranade@wandering.shop
                      wrote last edited by
                      #98

                      @glyph @SnoopJ I've mostly abstained from trying these tools in the process of criticism on the basis of ethical concerns, but secondarily — I know that the makers of these products are trying to trick me into thinking that they're something they're not. I don't trust my own cognition enough to keep putting it into danger, I guess?

                      glyph@mastodon.socialG snoopj@hachyderm.ioS 2 Replies Last reply
                      0
                      • xgranade@wandering.shopX xgranade@wandering.shop

                        @glyph @SnoopJ I've mostly abstained from trying these tools in the process of criticism on the basis of ethical concerns, but secondarily — I know that the makers of these products are trying to trick me into thinking that they're something they're not. I don't trust my own cognition enough to keep putting it into danger, I guess?

                        glyph@mastodon.socialG This user is from outside of this forum
                        glyph@mastodon.socialG This user is from outside of this forum
                        glyph@mastodon.social
                        wrote last edited by
                        #99

                        @xgranade @SnoopJ awareness of metacognitive distortions seems to be extremely predictive of llm skepticism, yes

                        xgranade@wandering.shopX 1 Reply Last reply
                        0
                        • snoopj@hachyderm.ioS snoopj@hachyderm.io

                          @glyph the little death that brings obliteration, perhaps

                          snoopj@hachyderm.ioS This user is from outside of this forum
                          snoopj@hachyderm.ioS This user is from outside of this forum
                          snoopj@hachyderm.io
                          wrote last edited by
                          #100

                          @glyph I can see how people, even experts, getting themselves into the scenario of "I have absolutely no idea what this code is doing or how it is organized and the only way out is by using the model harder"

                          (of course, the example of this most immediately visible to me should be a slam dunk: "oh jeez I forgot to tell it to write tests from the start" → "okay why can't you do that now?" → [anxiety noises])

                          1 Reply Last reply
                          0
                          • glyph@mastodon.socialG glyph@mastodon.social

                            @xgranade @SnoopJ awareness of metacognitive distortions seems to be extremely predictive of llm skepticism, yes

                            xgranade@wandering.shopX This user is from outside of this forum
                            xgranade@wandering.shopX This user is from outside of this forum
                            xgranade@wandering.shop
                            wrote last edited by
                            #101

                            @glyph @SnoopJ I don't think I'm an especial outlier on addiction susceptibility, in either direction. At most, I think I can say that most propaganda is designed for people who think fairly differently than I do... if someone is aiming more specifically at people like me, I don't think I'm immune.

                            I've seen far too many people who are at least vaguely similar to me come completely unglued on contact with these things, and I don't trust that I can resist that.

                            1 Reply Last reply
                            0
                            • xgranade@wandering.shopX xgranade@wandering.shop

                              @glyph @SnoopJ I've mostly abstained from trying these tools in the process of criticism on the basis of ethical concerns, but secondarily — I know that the makers of these products are trying to trick me into thinking that they're something they're not. I don't trust my own cognition enough to keep putting it into danger, I guess?

                              snoopj@hachyderm.ioS This user is from outside of this forum
                              snoopj@hachyderm.ioS This user is from outside of this forum
                              snoopj@hachyderm.io
                              wrote last edited by
                              #102

                              @xgranade @glyph yea, I had some thoughts about how bad this probably is for *my neurotype* (wtfever that is) specifically.

                              I am trying to "set those things aside" (ugh, those words taste bad) and approach this from a steel-man perspective, and answer: is it fit for purpose at the most mercenary level?

                              But I had not realized until I started using it just how awful the regular pauses in my work would be. Sure, I could turn my attention to something else while the model spins, but I know where *that* road goes and it's nowhere good.

                              And otherwise, I'm just sitting there waiting for my next cleanup on Aisle 3, and the fatigue set in quickly.

                              snoopj@hachyderm.ioS 1 Reply Last reply
                              0
                              • snoopj@hachyderm.ioS snoopj@hachyderm.io

                                @xgranade @glyph yea, I had some thoughts about how bad this probably is for *my neurotype* (wtfever that is) specifically.

                                I am trying to "set those things aside" (ugh, those words taste bad) and approach this from a steel-man perspective, and answer: is it fit for purpose at the most mercenary level?

                                But I had not realized until I started using it just how awful the regular pauses in my work would be. Sure, I could turn my attention to something else while the model spins, but I know where *that* road goes and it's nowhere good.

                                And otherwise, I'm just sitting there waiting for my next cleanup on Aisle 3, and the fatigue set in quickly.

                                snoopj@hachyderm.ioS This user is from outside of this forum
                                snoopj@hachyderm.ioS This user is from outside of this forum
                                snoopj@hachyderm.io
                                wrote last edited by
                                #103

                                @xgranade @glyph I'm almost relieved that it consumed as much compute as it did and *still* made mistakes on what was relatively still quite a straightforward task, even if it did involve some JSONSchema fancy footwork, and even if it did surface some issues.

                                Based on today's experimentation, the juice does not seem worth the squeeze. I have more evaluation to do in order to have my ducks in a row for when the Enthusiasts in my life come calling asking why I feel the way I do about this, but yea. Unsettling experience so far.

                                1 Reply Last reply
                                0
                                • snoopj@hachyderm.ioS snoopj@hachyderm.io

                                  But it was like a breath of fresh air to finish with the schema work, then point `datamodel-code-generator` (a deterministic tool) at those files and generate corresponding Pydantic model code for the schema.

                                  Not only was that Python code *much* easier to review (not the model's fault, JSONSchema is just very difficult for me to read) but it was just… so nice to be touching a tool whose behavior I can rely on. I didn't need to check its work very closely at all, because I know that it's applying a fixed set of well-characterized rules to generate those models from the (equally well-defined) schema.

                                  😮‍💨

                                  ancoghlan@mastodon.socialA This user is from outside of this forum
                                  ancoghlan@mastodon.socialA This user is from outside of this forum
                                  ancoghlan@mastodon.social
                                  wrote last edited by
                                  #104

                                  @SnoopJ I like datamodel-code-generator, but did have a few utterly baffling sessions with it that turned out to be due to a missing deepcopy call: https://github.com/koxudaxi/datamodel-code-generator/pull/2215

                                  (I shudder to think what nonsense an LLM might spew if it hit a tool bug like that)

                                  snoopj@hachyderm.ioS 1 Reply Last reply
                                  0
                                  • ancoghlan@mastodon.socialA ancoghlan@mastodon.social

                                    @SnoopJ I like datamodel-code-generator, but did have a few utterly baffling sessions with it that turned out to be due to a missing deepcopy call: https://github.com/koxudaxi/datamodel-code-generator/pull/2215

                                    (I shudder to think what nonsense an LLM might spew if it hit a tool bug like that)

                                    snoopj@hachyderm.ioS This user is from outside of this forum
                                    snoopj@hachyderm.ioS This user is from outside of this forum
                                    snoopj@hachyderm.io
                                    wrote last edited by
                                    #105

                                    @ancoghlan I imagine you've put it to much heavier-duty use than I am, but good to know!

                                    I also shudder to think. I suppose I should find a suitable bug from my own recent past and see how long is the path from the initial report to proper characterization of the issue (if we get there at all)

                                    1 Reply Last reply
                                    0
                                    • snoopj@hachyderm.ioS snoopj@hachyderm.io

                                      Coming for your job

                                      dpnash@c.imD This user is from outside of this forum
                                      dpnash@c.imD This user is from outside of this forum
                                      dpnash@c.im
                                      wrote last edited by
                                      #106

                                      @SnoopJ

                                      Yikes. Just the first line of output requires both March 3rd and March 5th to be Tuesday—at least, if you remember obscure facts like “Tuesday is the day after Monday” and “there are 7 days in a week.”

                                      Truly this application has a dizzying intellect and can be completely trusted with any other modular arithmetic it might stumble across.

                                      1 Reply Last reply
                                      0
                                      • snoopj@hachyderm.ioS snoopj@hachyderm.io

                                        But it was like a breath of fresh air to finish with the schema work, then point `datamodel-code-generator` (a deterministic tool) at those files and generate corresponding Pydantic model code for the schema.

                                        Not only was that Python code *much* easier to review (not the model's fault, JSONSchema is just very difficult for me to read) but it was just… so nice to be touching a tool whose behavior I can rely on. I didn't need to check its work very closely at all, because I know that it's applying a fixed set of well-characterized rules to generate those models from the (equally well-defined) schema.

                                        😮‍💨

                                        snoopj@hachyderm.ioS This user is from outside of this forum
                                        snoopj@hachyderm.ioS This user is from outside of this forum
                                        snoopj@hachyderm.io
                                        wrote last edited by
                                        #107

                                        gave the bullshit machine another gentle pitch, fed it the report of new bug and asked for an explanation

                                        it did correctly point to where the originating flaw was (`data=…` instead of `json=…` in `requests`), and from where I hit pause and explored possible explanations for why that suddenly mattered (FastAPI 0.132 has a breaking change associated with the "wrong" `Content-Type` header)

                                        then, already knowing what the problem was, asked for an *explanation* rather than "oh look at this code". took several cycles of incorrect confabulation including very explicit hints ("I am sure the Pydantic version has not changed." "FastAPI is not bounded above, check the release notes") to get to an explanation that could be called correct.

                                        I don't know how to evaluate how much faster that got me through the fog-of-war that is Pydantic's absolutely terrible error reporting, but I do know that the number of potential pitfalls on the far side of that is… not something that inspires faith.

                                        snoopj@hachyderm.ioS delta_vee@mstdn.caD 2 Replies Last reply
                                        0
                                        • snoopj@hachyderm.ioS snoopj@hachyderm.io

                                          gave the bullshit machine another gentle pitch, fed it the report of new bug and asked for an explanation

                                          it did correctly point to where the originating flaw was (`data=…` instead of `json=…` in `requests`), and from where I hit pause and explored possible explanations for why that suddenly mattered (FastAPI 0.132 has a breaking change associated with the "wrong" `Content-Type` header)

                                          then, already knowing what the problem was, asked for an *explanation* rather than "oh look at this code". took several cycles of incorrect confabulation including very explicit hints ("I am sure the Pydantic version has not changed." "FastAPI is not bounded above, check the release notes") to get to an explanation that could be called correct.

                                          I don't know how to evaluate how much faster that got me through the fog-of-war that is Pydantic's absolutely terrible error reporting, but I do know that the number of potential pitfalls on the far side of that is… not something that inspires faith.

                                          snoopj@hachyderm.ioS This user is from outside of this forum
                                          snoopj@hachyderm.ioS This user is from outside of this forum
                                          snoopj@hachyderm.io
                                          wrote last edited by
                                          #108

                                          this kind of use at least puts *some* limitation on the blast radius of the use

                                          this experiment and another (a descriptive task mentioned up-thread) convinces me that there are still many pitfalls when using such a model to guide this kind of exploratory work

                                          but at least in that case, the user of the tool is the one who is having their time wasted the most, and bullshit does not as easily make it into the codebase.

                                          assuming the tool has been instructed not to generate any and the user has not directly subverted that, anyway. which is of course a big assumption, sufficient Enthusiasm will (does) clearly lead to this kind of subversion and lying about it

                                          snoopj@hachyderm.ioS 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups