Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. hm https://github.com/bluesky-social/social-app/blob/main/CLAUDE.md

hm https://github.com/bluesky-social/social-app/blob/main/CLAUDE.md

Scheduled Pinned Locked Moved Uncategorized
108 Posts 55 Posters 289 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • corujosilva@fedi.latC corujosilva@fedi.lat

    @cwebber before* any judgement on whethe it is a good thing or not, it was expected, tbh. it is very much on brand from their team.
    they always had the "tech enthusiast" ethos

    *just before.

    davidgerard@circumstances.runD This user is from outside of this forum
    davidgerard@circumstances.runD This user is from outside of this forum
    davidgerard@circumstances.run
    wrote last edited by
    #73

    @corujosilva @cwebber the team started as coiners, so they were very hot for AI

    1 Reply Last reply
    0
    • november@chaosfem.twN november@chaosfem.tw

      @cwebber Ah yes, LLMs are finally good now, this is probably the fourth or fifth time I've heard it and at this point it's like the boy who cried wolf, I'm not even going to bother testing out the LLMs of today to see what they get wrong, I'm just not going to believe their advocates

      dzwiedziu@mastodon.socialD This user is from outside of this forum
      dzwiedziu@mastodon.socialD This user is from outside of this forum
      dzwiedziu@mastodon.social
      wrote last edited by
      #74

      @november
      Wait? LLMs are good now? How nobody told me, except swaths of… techbros… …oh wait…

      @cwebber

      1 Reply Last reply
      0
      • cwebber@social.coopC cwebber@social.coop

        hm https://github.com/bluesky-social/social-app/blob/main/CLAUDE.md

        mkljczk@pl.fediverse.plM This user is from outside of this forum
        mkljczk@pl.fediverse.plM This user is from outside of this forum
        mkljczk@pl.fediverse.pl
        wrote last edited by
        #75
        @cwebber

        RT: https://pl.fediverse.pl/objects/5174b877-5def-4a70-b561-cefdd62c9fa9
        mkljczk@pl.fediverse.plM 1 Reply Last reply
        0
        • erincandescent@akko.erincandescent.netE erincandescent@akko.erincandescent.net
          @cwebber @res260 I feel like there's always been a lot of software development that isn't craft but it's just shuffling bits around.

          I don't really know how to feel about a lot of it these days. I've played around with some of the tools for work and there's certainly a lot of areas where they can write basically the same code that I would have done with less tedium, and by some metrics they do a better job (mostly things that are good practice but I couldn't be bothered).

          Is that abandoning craft or careful allocation of executive function? I don't know.

          I definitely think these things aren't going away. The bubble will pop, it'll maybe kill the big AI companies, people will stop shoving chat bots everywhere, but I don't see any way that LLMs don't remain a fact of life, and I don't know what the long term implications are of this
          sitcom_nemesis@tech.lgbtS This user is from outside of this forum
          sitcom_nemesis@tech.lgbtS This user is from outside of this forum
          sitcom_nemesis@tech.lgbt
          wrote last edited by
          #76

          @erincandescent @res260 @cwebber I am no fan of tedium for tedium's sake... but why specifically slop machines, which are notoriously unreliable, to solve this problem, aside from all the money that got poured into this technology? Could the same money have been used to develop languages and frameworks with sensible defaults and configurations, thereby eliminating (or vastly reducing) the need for tedium?

          erincandescent@akko.erincandescent.netE 1 Reply Last reply
          0
          • benjamineskola@hachyderm.ioB benjamineskola@hachyderm.io

            @erincandescent @res260 @cwebber @airtower

            'modulo institutional knowledge' is doing a lot of heavy lifting there since that's half the problem with LLM usage

            and the other half of the problem is the assumption that an LLM will produce identical code

            so I don't think there's a useful discussion to be had if those are your assumptions

            erincandescent@akko.erincandescent.netE This user is from outside of this forum
            erincandescent@akko.erincandescent.netE This user is from outside of this forum
            erincandescent@akko.erincandescent.net
            wrote last edited by
            #77

            @benjamineskola @res260 @cwebber @airtower Look, I don’t think we’re talking about (original definition) vibe coding here, where nobody is looking at the output. We’re talking about cases where there’s a human in the loop.

            If the tool is generating garbage code and the human is accepting it, that’s a human problem more than a tool problem.

            I start from this assumption because we assume the human is competent and has taste. I assume they are not just letting the tool run wild on the codebase and make a mess.

            There are issues and questions around institutional knowledge (if the human isn’t exploring the codebase in the same way, how much are they learning? how much do you pickup through review vs implementation?) but even then I’d argue that one of the primary criterions with regards to maintainability is how hard it is for a newcomer to pick something up and work on it.

            benjamineskola@hachyderm.ioB 1 Reply Last reply
            0
            • november@chaosfem.twN november@chaosfem.tw

              @cwebber Ah yes, LLMs are finally good now, this is probably the fourth or fifth time I've heard it and at this point it's like the boy who cried wolf, I'm not even going to bother testing out the LLMs of today to see what they get wrong, I'm just not going to believe their advocates

              lebramor@piaille.frL This user is from outside of this forum
              lebramor@piaille.frL This user is from outside of this forum
              lebramor@piaille.fr
              wrote last edited by
              #78

              @november @cwebber but at one point the boy cried wolf for a real reason.

              1 Reply Last reply
              0
              • erincandescent@akko.erincandescent.netE erincandescent@akko.erincandescent.net

                @benjamineskola @res260 @cwebber @airtower Look, I don’t think we’re talking about (original definition) vibe coding here, where nobody is looking at the output. We’re talking about cases where there’s a human in the loop.

                If the tool is generating garbage code and the human is accepting it, that’s a human problem more than a tool problem.

                I start from this assumption because we assume the human is competent and has taste. I assume they are not just letting the tool run wild on the codebase and make a mess.

                There are issues and questions around institutional knowledge (if the human isn’t exploring the codebase in the same way, how much are they learning? how much do you pickup through review vs implementation?) but even then I’d argue that one of the primary criterions with regards to maintainability is how hard it is for a newcomer to pick something up and work on it.

                benjamineskola@hachyderm.ioB This user is from outside of this forum
                benjamineskola@hachyderm.ioB This user is from outside of this forum
                benjamineskola@hachyderm.io
                wrote last edited by
                #79

                @erincandescent @res260 @cwebber @airtower Except there is a huge problem with people actually just not looking at the code being generated. The wave of slop PRs inundating many open-source projects recently, for example.

                People keep saying 'of course there is a human in the loop' but it seems increasingly clear to me that nobody is actually bothering to be the human in the loop themselves.

                (Edit: but also, even when people are well-intentioned, I think the LLM-based process just makes it much harder to ensure quality than actually writing the code oneself.)

                And yes, this is a human problem, it's all a human problem. But that's like saying 'guns don't kill people, people do'. True, but, the tool clearly exacerbates the problem.

                As for your final paragraph I don't remotely see why you think LLMs solve this problem either.

                erincandescent@akko.erincandescent.netE 1 Reply Last reply
                0
                • sitcom_nemesis@tech.lgbtS sitcom_nemesis@tech.lgbt

                  @erincandescent @res260 @cwebber I am no fan of tedium for tedium's sake... but why specifically slop machines, which are notoriously unreliable, to solve this problem, aside from all the money that got poured into this technology? Could the same money have been used to develop languages and frameworks with sensible defaults and configurations, thereby eliminating (or vastly reducing) the need for tedium?

                  erincandescent@akko.erincandescent.netE This user is from outside of this forum
                  erincandescent@akko.erincandescent.netE This user is from outside of this forum
                  erincandescent@akko.erincandescent.net
                  wrote last edited by
                  #80

                  @sitcom_nemesis @res260 @cwebber I think there’s a spectrum

                  There’s code we keep repeating in broadly the exact same structure, just with different details fileld in. That’s boilerplate.

                  There’s code that’s unique and creative and requires thought. That’s “the meat of the problem”.

                  But there’s lots of stuff in the middle where it’s not quite creative, doesn’t really require thought, but either because of domain requirements, accidents of history, or just because you’re gluing two libraries together that hadn’t ever seen each other, is too irregular to really code generate but is not actually interesting.

                  1 Reply Last reply
                  0
                  • mkljczk@pl.fediverse.plM mkljczk@pl.fediverse.pl
                    @cwebber

                    RT: https://pl.fediverse.pl/objects/5174b877-5def-4a70-b561-cefdd62c9fa9
                    mkljczk@pl.fediverse.plM This user is from outside of this forum
                    mkljczk@pl.fediverse.plM This user is from outside of this forum
                    mkljczk@pl.fediverse.pl
                    wrote last edited by
                    #81
                    @cwebber But honestly bsky web UI felt like a one-shoted ‘vibe-coded’ Twitter clone even before it actually became one.
                    1 Reply Last reply
                    0
                    • benjamineskola@hachyderm.ioB benjamineskola@hachyderm.io

                      @erincandescent @res260 @cwebber @airtower

                      'modulo institutional knowledge' is doing a lot of heavy lifting there since that's half the problem with LLM usage

                      and the other half of the problem is the assumption that an LLM will produce identical code

                      so I don't think there's a useful discussion to be had if those are your assumptions

                      airtower@woem.menA This user is from outside of this forum
                      airtower@woem.menA This user is from outside of this forum
                      airtower@woem.men
                      wrote last edited by
                      #82

                      @benjamineskola@hachyderm.io @erincandescent@akko.erincandescent.net @res260@infosec.exchange @cwebber@social.coop Yeah, this. If I'm looking at a new tool/library to possibly use (and not do a hard fork on), a key question is: Are there people who understand and care maintaining this thing? Because if there aren't, eventually "hard fork" or "don't use it" will probably be my only choices.

                      And using LLMs to generate code points towards "no" (or at least "not much") for both understanding and caring. If someone skilled is actually putting in the effort to edit LLM output until it is no worse than what they would've written themselves (point for care at least), chances are it would've been faster (let alone other effects) to just do that.

                      erincandescent@akko.erincandescent.netE 1 Reply Last reply
                      0
                      • benjamineskola@hachyderm.ioB benjamineskola@hachyderm.io

                        @erincandescent @res260 @cwebber @airtower Except there is a huge problem with people actually just not looking at the code being generated. The wave of slop PRs inundating many open-source projects recently, for example.

                        People keep saying 'of course there is a human in the loop' but it seems increasingly clear to me that nobody is actually bothering to be the human in the loop themselves.

                        (Edit: but also, even when people are well-intentioned, I think the LLM-based process just makes it much harder to ensure quality than actually writing the code oneself.)

                        And yes, this is a human problem, it's all a human problem. But that's like saying 'guns don't kill people, people do'. True, but, the tool clearly exacerbates the problem.

                        As for your final paragraph I don't remotely see why you think LLMs solve this problem either.

                        erincandescent@akko.erincandescent.netE This user is from outside of this forum
                        erincandescent@akko.erincandescent.netE This user is from outside of this forum
                        erincandescent@akko.erincandescent.net
                        wrote last edited by
                        #83

                        @benjamineskola @res260 @cwebber @airtower

                        Except there is a huge problem with people actually just not looking at the code being generated. The wave of slop PRs inundating many open-source projects recently, for example.

                        People keep saying ‘of course there is a human in the loop’ but it seems increasingly clear to me that nobody is actually bothering to be the human in the loop themselves.

                        I know these are problems, but you’re moving the topic of conversation. There have always been bad developers with bad practices shovling crappy code over the fence. LLMs have made this easier and it sucks but it’s not new.

                        And yes, this is a human problem, it’s all a human problem. But that’s like saying ‘guns don’t kill people, people do’. True, but, the tool clearly exacerbates the problem.

                        Sure, but lazy/careless people use tool to produce bad results is not a unique problem. It’s very easy with a power drill to make messy holes, but we arent’ forcing everyone to use hand drills.

                        Saying using these tools results in necessarily bad output is just not backed up by available evidence.

                        I don’t pretend they’re perfect and I don’t pretend there aren’t problems. What I sense is that they’re not going away and are going to become and remain routine parts of toolboxes long into the future.

                        benjamineskola@hachyderm.ioB 1 Reply Last reply
                        0
                        • erincandescent@akko.erincandescent.netE erincandescent@akko.erincandescent.net

                          @benjamineskola @res260 @cwebber @airtower

                          Except there is a huge problem with people actually just not looking at the code being generated. The wave of slop PRs inundating many open-source projects recently, for example.

                          People keep saying ‘of course there is a human in the loop’ but it seems increasingly clear to me that nobody is actually bothering to be the human in the loop themselves.

                          I know these are problems, but you’re moving the topic of conversation. There have always been bad developers with bad practices shovling crappy code over the fence. LLMs have made this easier and it sucks but it’s not new.

                          And yes, this is a human problem, it’s all a human problem. But that’s like saying ‘guns don’t kill people, people do’. True, but, the tool clearly exacerbates the problem.

                          Sure, but lazy/careless people use tool to produce bad results is not a unique problem. It’s very easy with a power drill to make messy holes, but we arent’ forcing everyone to use hand drills.

                          Saying using these tools results in necessarily bad output is just not backed up by available evidence.

                          I don’t pretend they’re perfect and I don’t pretend there aren’t problems. What I sense is that they’re not going away and are going to become and remain routine parts of toolboxes long into the future.

                          benjamineskola@hachyderm.ioB This user is from outside of this forum
                          benjamineskola@hachyderm.ioB This user is from outside of this forum
                          benjamineskola@hachyderm.io
                          wrote last edited by
                          #84

                          @erincandescent @res260 @cwebber @airtower > LLMs have made this easier and it sucks but it’s not new.

                          So why would we want to make it worse?

                          > Saying using these tools results in necessarily bad output is just not backed up by available evidence.

                          Every output I've seen from these things has been, at best, no better than a human would have done. And that's being generous.

                          > What I sense is that they’re not going away and are going to become and remain routine parts of toolboxes long into the future.

                          This is a self-fulfilling prophecy. Of course they won't go away if people insist on defending them.

                          1 Reply Last reply
                          0
                          • airtower@woem.menA airtower@woem.men

                            @benjamineskola@hachyderm.io @erincandescent@akko.erincandescent.net @res260@infosec.exchange @cwebber@social.coop Yeah, this. If I'm looking at a new tool/library to possibly use (and not do a hard fork on), a key question is: Are there people who understand and care maintaining this thing? Because if there aren't, eventually "hard fork" or "don't use it" will probably be my only choices.

                            And using LLMs to generate code points towards "no" (or at least "not much") for both understanding and caring. If someone skilled is actually putting in the effort to edit LLM output until it is no worse than what they would've written themselves (point for care at least), chances are it would've been faster (let alone other effects) to just do that.

                            erincandescent@akko.erincandescent.netE This user is from outside of this forum
                            erincandescent@akko.erincandescent.netE This user is from outside of this forum
                            erincandescent@akko.erincandescent.net
                            wrote last edited by
                            #85
                            @airtower @res260 @benjamineskola @cwebber I still see LLM related artifacts as a negative quality signal. There's lots of crap LLM aided code out there and there's lots of people slopping stuff together. The worst developers are disproportionality interested.

                            But I think there's a lot of stuff being written with LLM assistance these days where you'd not be able to tell
                            airtower@woem.menA benjamineskola@hachyderm.ioB 2 Replies Last reply
                            0
                            • erincandescent@akko.erincandescent.netE erincandescent@akko.erincandescent.net
                              @airtower @res260 @benjamineskola @cwebber I still see LLM related artifacts as a negative quality signal. There's lots of crap LLM aided code out there and there's lots of people slopping stuff together. The worst developers are disproportionality interested.

                              But I think there's a lot of stuff being written with LLM assistance these days where you'd not be able to tell
                              airtower@woem.menA This user is from outside of this forum
                              airtower@woem.menA This user is from outside of this forum
                              airtower@woem.men
                              wrote last edited by
                              #86

                              @erincandescent@akko.erincandescent.net @res260@infosec.exchange @benjamineskola@hachyderm.io @cwebber@social.coop That might be, but as I wrote in that case I doubt there's any benefit (like faster progress) to the developer (even looking at code only, ignoring all the harmful side effects of LLMs).

                              1 Reply Last reply
                              0
                              • erincandescent@akko.erincandescent.netE erincandescent@akko.erincandescent.net
                                @airtower @res260 @benjamineskola @cwebber I still see LLM related artifacts as a negative quality signal. There's lots of crap LLM aided code out there and there's lots of people slopping stuff together. The worst developers are disproportionality interested.

                                But I think there's a lot of stuff being written with LLM assistance these days where you'd not be able to tell
                                benjamineskola@hachyderm.ioB This user is from outside of this forum
                                benjamineskola@hachyderm.ioB This user is from outside of this forum
                                benjamineskola@hachyderm.io
                                wrote last edited by
                                #87

                                @erincandescent @res260 @cwebber @airtower Given that every output of LLMs that I've seen that is identifiable as such has been mediocre at best, why would I assume without any evidence that there's a significant quantity of LLM-generated code that's actually good?

                                "There's no evidence of it but it's definitely there" is unpersuasive.

                                And I've also found that people's evaluations of LLM-generated code quality is wildly out of step with my own evaluations, so I would not automatically assume that because someone says it's good that it's actually good.

                                And then, even if the code was of acceptable quality, the negative effects on the process (increased difficult of reviewing ↔ decreased institutional knowledge, among other things) count against it too.

                                (And all of this is setting aside the ethical issues, which in practice I don't think we should do anyway. Like, even if LLMs produced good output they'd be ethically indefensible, and even if they were ethically acceptable the results are so poor that why would you bother with them?)

                                1 Reply Last reply
                                0
                                • joeyh@sunbeam.cityJ joeyh@sunbeam.city

                                  @cwebber I found LLM generated code in vim today

                                  gnomon@mastodon.socialG This user is from outside of this forum
                                  gnomon@mastodon.socialG This user is from outside of this forum
                                  gnomon@mastodon.social
                                  wrote last edited by
                                  #88

                                  @joeyh @cwebber well that is mighty disappointing.

                                  benjamineskola@hachyderm.ioB 1 Reply Last reply
                                  0
                                  • gnomon@mastodon.socialG gnomon@mastodon.social

                                    @joeyh @cwebber well that is mighty disappointing.

                                    benjamineskola@hachyderm.ioB This user is from outside of this forum
                                    benjamineskola@hachyderm.ioB This user is from outside of this forum
                                    benjamineskola@hachyderm.io
                                    wrote last edited by
                                    #89

                                    @gnomon @joeyh @cwebber They were advertising some LLM stuff in the most recent release notes but it didn't sound like they were actually using it on Vim itself.

                                    On the other hand, anyone who cares enough to advertise LLMs is probably using them too.

                                    1 Reply Last reply
                                    0
                                    • cwebber@social.coopC cwebber@social.coop

                                      I have this suspicion that the ATproto stack, at least the stuff from Bluesky, is heading towards "majority-vibecoded" but that's mostly just from seeing a lot of posts from the Bluesky eng team rather than me having spent much time in the codebase

                                      Why is def hugely responsible for Bluesky/ATProto's design and if *he's* mostly letting Claude write 99% of his code, the rest of the eng team is likely to be heading in that direction too?

                                      cwebber@social.coopC This user is from outside of this forum
                                      cwebber@social.coopC This user is from outside of this forum
                                      cwebber@social.coop
                                      wrote last edited by
                                      #90

                                      Also https://bsky.app/profile/pfrazee.com/post/3meogr22l3k2d

                                      > A year ago, I thought LLMs were kind of neat but not that useful. I saw the code autocomplete and thought, meh.
                                      >
                                      >Last summer just flipped. I never ever thought I would see automated code generation like we see now.
                                      >
                                      > I know there’s baggage but you need to know the coders are being real about this

                                      cwebber@social.coopC aslakr@mastodon.socialA 2 Replies Last reply
                                      0
                                      • cwebber@social.coopC cwebber@social.coop

                                        Also https://bsky.app/profile/pfrazee.com/post/3meogr22l3k2d

                                        > A year ago, I thought LLMs were kind of neat but not that useful. I saw the code autocomplete and thought, meh.
                                        >
                                        >Last summer just flipped. I never ever thought I would see automated code generation like we see now.
                                        >
                                        > I know there’s baggage but you need to know the coders are being real about this

                                        cwebber@social.coopC This user is from outside of this forum
                                        cwebber@social.coopC This user is from outside of this forum
                                        cwebber@social.coop
                                        wrote last edited by
                                        #91

                                        Also https://bsky.app/profile/samuel.fm/post/3mbz27d6qnc2v

                                        eramdam@social.erambert.meE cwebber@social.coopC 2 Replies Last reply
                                        0
                                        • cwebber@social.coopC cwebber@social.coop

                                          Also https://bsky.app/profile/samuel.fm/post/3mbz27d6qnc2v

                                          eramdam@social.erambert.meE This user is from outside of this forum
                                          eramdam@social.erambert.meE This user is from outside of this forum
                                          eramdam@social.erambert.me
                                          wrote last edited by
                                          #92

                                          @cwebber yeah, least surprising Bluesky thing to do

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups