Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink).

I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink).

Scheduled Pinned Locked Moved Uncategorized
theclimatebrinkaicoding
56 Posts 19 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • arnebab@rollenspiel.socialA arnebab@rollenspiel.social

    @Ruth_Mottram One experiment I did was to turn a text I wrote years ago into a scientific paper in economics.

    It took two hours and reached a quality that I (physicist, not from economics) could not have distinguished it from a real paper.

    AI causes the form to be easier to repeat, so we can no longer trust the form of scientific writing to be a hint that people actually have scientific education.

    And that is a huge risk.
    @benjamingeer @hausfath @UlrikeHahn

    arnebab@rollenspiel.socialA This user is from outside of this forum
    arnebab@rollenspiel.socialA This user is from outside of this forum
    arnebab@rollenspiel.social
    wrote last edited by
    #31

    @Ruth_Mottram though my main gripe with us as human society is that we’re spending more than 400 billion dollars a year to build error-prone general pattern recognition and reproduction while finding maybe 100 problems where it brings big benefits -- that would each require less than 10 million dollars to solve.

    Why don’t we have solutions for those tasks already?

    Why is matplotlib mostly written by some folks in their spare time while it has tons of value?
    @benjamingeer @hausfath @UlrikeHahn

    1 Reply Last reply
    0
    • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

      @benjamingeer @Ruth_Mottram @hausfath sometimes replies here leave me speechless…

      benjamingeer@piaille.frB This user is from outside of this forum
      benjamingeer@piaille.frB This user is from outside of this forum
      benjamingeer@piaille.fr
      wrote last edited by
      #32

      @UlrikeHahn What is the "good" that you want your students to produce? The thing that has real value? Is it essays or learning? Perhaps students are using LLMs to write essays because they mistakenly believe that the essay is an end in itself, rather than a means to an end. As somebody said, sometimes it makes sense to have someone cook your meal for you, but it never makes sense to have someone eat your meal for you. @Ruth_Mottram @hausfath

      ulrikehahn@fediscience.orgU 1 Reply Last reply
      0
      • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

        I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

        Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
        Because this kind of performance changes everything and we can't afford to lag behind...
        #AIcoding #ML

        Edit: here is the climate brink post I mentioned

        Link Preview Image
        The AI-Augmented Scientist

        The promise and pitfalls of using AI tools to boost my capabilities as a scientist

        favicon

        (www.theclimatebrink.com)

        karolina@fediscience.orgK This user is from outside of this forum
        karolina@fediscience.orgK This user is from outside of this forum
        karolina@fediscience.org
        wrote last edited by
        #33

        Do people actually read the code Claude runs and how it differs from what Claude gives as an output?

        1 Reply Last reply
        0
        • benjamingeer@piaille.frB benjamingeer@piaille.fr

          @UlrikeHahn What is the "good" that you want your students to produce? The thing that has real value? Is it essays or learning? Perhaps students are using LLMs to write essays because they mistakenly believe that the essay is an end in itself, rather than a means to an end. As somebody said, sometimes it makes sense to have someone cook your meal for you, but it never makes sense to have someone eat your meal for you. @Ruth_Mottram @hausfath

          ulrikehahn@fediscience.orgU This user is from outside of this forum
          ulrikehahn@fediscience.orgU This user is from outside of this forum
          ulrikehahn@fediscience.org
          wrote last edited by
          #34

          @benjamingeer @Ruth_Mottram @hausfath Benjamin, maybe just reread the previous post of yours and ask yourself “what in this post am I saying that could possibly be new to the person I am addressing?”…and then see where that leads you

          benjamingeer@piaille.frB 1 Reply Last reply
          0
          • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

            @benjamingeer @Ruth_Mottram @hausfath Benjamin, maybe just reread the previous post of yours and ask yourself “what in this post am I saying that could possibly be new to the person I am addressing?”…and then see where that leads you

            benjamingeer@piaille.frB This user is from outside of this forum
            benjamingeer@piaille.frB This user is from outside of this forum
            benjamingeer@piaille.fr
            wrote last edited by
            #35

            @UlrikeHahn It would surprise me if anything I said was new to you. What surprised me was that you described the production of counterfeit goods as productivity. @Ruth_Mottram @hausfath

            ulrikehahn@fediscience.orgU 1 Reply Last reply
            0
            • benjamingeer@piaille.frB benjamingeer@piaille.fr

              @UlrikeHahn It would surprise me if anything I said was new to you. What surprised me was that you described the production of counterfeit goods as productivity. @Ruth_Mottram @hausfath

              ulrikehahn@fediscience.orgU This user is from outside of this forum
              ulrikehahn@fediscience.orgU This user is from outside of this forum
              ulrikehahn@fediscience.org
              wrote last edited by
              #36

              @benjamingeer @Ruth_Mottram @hausfath maybe that should be a clue that you are somehow missing the intended point?

              benjamingeer@piaille.frB 1 Reply Last reply
              0
              • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                @benjamingeer @Ruth_Mottram @hausfath maybe that should be a clue that you are somehow missing the intended point?

                benjamingeer@piaille.frB This user is from outside of this forum
                benjamingeer@piaille.frB This user is from outside of this forum
                benjamingeer@piaille.fr
                wrote last edited by
                #37

                @UlrikeHahn The original question was whether LLM coding assistants would make scientists more productive. It sounded like you were arguing that they would, since LLMs are not just hype, as evidenced by their efficiency in producing fake course work, etc. Were you being ironic? @Ruth_Mottram @hausfath

                ulrikehahn@fediscience.orgU 1 Reply Last reply
                0
                • arnebab@rollenspiel.socialA arnebab@rollenspiel.social

                  @Ruth_Mottram One experiment I did was to turn a text I wrote years ago into a scientific paper in economics.

                  It took two hours and reached a quality that I (physicist, not from economics) could not have distinguished it from a real paper.

                  AI causes the form to be easier to repeat, so we can no longer trust the form of scientific writing to be a hint that people actually have scientific education.

                  And that is a huge risk.
                  @benjamingeer @hausfath @UlrikeHahn

                  arnebab@rollenspiel.socialA This user is from outside of this forum
                  arnebab@rollenspiel.socialA This user is from outside of this forum
                  arnebab@rollenspiel.social
                  wrote last edited by
                  #38

                  @Ruth_Mottram when you use AI to transform your content from one form to another, parts of the content usually associated with the target form creep into your content.

                  This can be as bad as turning "agriculture that needs less antibiotics, because animals stay healthier" into "agriculture without antibiotics" (so sick animals suffer needlessly).

                  Because AI does not differentiate between content and form.
                  @benjamingeer @hausfath @UlrikeHahn

                  1 Reply Last reply
                  0
                  • benjamingeer@piaille.frB benjamingeer@piaille.fr

                    @UlrikeHahn The original question was whether LLM coding assistants would make scientists more productive. It sounded like you were arguing that they would, since LLMs are not just hype, as evidenced by their efficiency in producing fake course work, etc. Were you being ironic? @Ruth_Mottram @hausfath

                    ulrikehahn@fediscience.orgU This user is from outside of this forum
                    ulrikehahn@fediscience.orgU This user is from outside of this forum
                    ulrikehahn@fediscience.org
                    wrote last edited by
                    #39

                    @benjamingeer @Ruth_Mottram @hausfath I will leave that to you to puzzle out and now stop bombarding Ruth’s thread….

                    1 Reply Last reply
                    0
                    • R relay@relay.an.exchange shared this topic
                    • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                      I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

                      Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
                      Because this kind of performance changes everything and we can't afford to lag behind...
                      #AIcoding #ML

                      Edit: here is the climate brink post I mentioned

                      Link Preview Image
                      The AI-Augmented Scientist

                      The promise and pitfalls of using AI tools to boost my capabilities as a scientist

                      favicon

                      (www.theclimatebrink.com)

                      osma@mas.toO This user is from outside of this forum
                      osma@mas.toO This user is from outside of this forum
                      osma@mas.to
                      wrote last edited by
                      #40

                      On pure software side: 10 years ago playing with the first gen Raspberry Pi camera, I realized its relatively exotic video interface could be leveraged to do motion detection with extremely low CPU usage.

                      Those interfaces have since changed and the same approach no longer works. So a few months ago I decided to try an experiment: could OpenCode make a new version, compatible with the latest hardware and interfaces? 1/2
                      @Ruth_Mottram @hausfath

                      osma@mas.toO 1 Reply Last reply
                      0
                      • osma@mas.toO osma@mas.to

                        On pure software side: 10 years ago playing with the first gen Raspberry Pi camera, I realized its relatively exotic video interface could be leveraged to do motion detection with extremely low CPU usage.

                        Those interfaces have since changed and the same approach no longer works. So a few months ago I decided to try an experiment: could OpenCode make a new version, compatible with the latest hardware and interfaces? 1/2
                        @Ruth_Mottram @hausfath

                        osma@mas.toO This user is from outside of this forum
                        osma@mas.toO This user is from outside of this forum
                        osma@mas.to
                        wrote last edited by
                        #41

                        The planning stage worked like magic. It generated a plan which detailed why the old code doesn't work, listed all new new solutions, and outlined a plan of conversion.

                        It all fell apart moving to implementation though. Spinning in circles it ended up producing a completely unworkable resemblance of code that didn't even have hope of working.

                        What looked excitingly plausible for a forward port turned out a dead end. 2/2
                        @Ruth_Mottram @hausfath

                        osma@mas.toO 1 Reply Last reply
                        0
                        • osma@mas.toO osma@mas.to

                          The planning stage worked like magic. It generated a plan which detailed why the old code doesn't work, listed all new new solutions, and outlined a plan of conversion.

                          It all fell apart moving to implementation though. Spinning in circles it ended up producing a completely unworkable resemblance of code that didn't even have hope of working.

                          What looked excitingly plausible for a forward port turned out a dead end. 2/2
                          @Ruth_Mottram @hausfath

                          osma@mas.toO This user is from outside of this forum
                          osma@mas.toO This user is from outside of this forum
                          osma@mas.to
                          wrote last edited by
                          #42

                          Since I didn't spend the time to try and implement the plan by hand, I don't know if it was feasible, just that it did look plausible at first.

                          And that I think is the major issue with all LLMs. The artifacts look plausible, entirely regardless of whether they're factually correct. 3/2
                          @Ruth_Mottram @hausfath

                          1 Reply Last reply
                          0
                          • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                            I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

                            Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
                            Because this kind of performance changes everything and we can't afford to lag behind...
                            #AIcoding #ML

                            Edit: here is the climate brink post I mentioned

                            Link Preview Image
                            The AI-Augmented Scientist

                            The promise and pitfalls of using AI tools to boost my capabilities as a scientist

                            favicon

                            (www.theclimatebrink.com)

                            T This user is from outside of this forum
                            T This user is from outside of this forum
                            tkissing@mastodon.social
                            wrote last edited by
                            #43

                            @Ruth_Mottram@fediscience.org @hausfath I played roulette once, putting $5 on a number and won. I didn't suggest that everyone I know should quit their jobs and just bet on that number for a living.

                            LLMs are autocomplete on cocaine. Yes, sometimes they'll spit out something useful, but often times they don't and the more we use them, the more we lose the ability to tell the good from the bad.

                            The best Europe can do is to invest in people.

                            1 Reply Last reply
                            0
                            • arnebab@rollenspiel.socialA arnebab@rollenspiel.social

                              @benjamingeer But, just to make it clear: that code which was 100x slower than it could have been, was still correct.

                              It was slow, but it did very complex tasks correctly.
                              @Ruth_Mottram @hausfath @UlrikeHahn

                              arnebab@rollenspiel.socialA This user is from outside of this forum
                              arnebab@rollenspiel.socialA This user is from outside of this forum
                              arnebab@rollenspiel.social
                              wrote last edited by
                              #44

                              @benjamingeer Therefore I’d rather compare LLMs to using statistical methods without understanding them.

                              That’s already widespread and I expect that with LLMs it will get worse.
                              @Ruth_Mottram @hausfath @UlrikeHahn

                              1 Reply Last reply
                              0
                              • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                                I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

                                Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
                                Because this kind of performance changes everything and we can't afford to lag behind...
                                #AIcoding #ML

                                Edit: here is the climate brink post I mentioned

                                Link Preview Image
                                The AI-Augmented Scientist

                                The promise and pitfalls of using AI tools to boost my capabilities as a scientist

                                favicon

                                (www.theclimatebrink.com)

                                1 This user is from outside of this forum
                                1 This user is from outside of this forum
                                1337@techhub.social
                                wrote last edited by
                                #45

                                @Ruth_Mottram @hausfath This seems like a *really* bad idea. I'm a software engineer and not a scientist, but I believe I've heard there's already a fairly big problem in the sciences with software bugs producing misleading results. I imagine using AI to write code could make this much worse. IMO, the extra time that would've been spent coding everything would not have been wasted. Coding it yourself gives you more time to think about what you're typing and gain a more complete understanding of your code and the libraries you're using; giving you more time and insight to spot bugs or otherwise wrong or less than optimal ways of doing things. If one did a thorough review of the AI generated code to ensure it was correct, I'd guess it take at least the same amount of time. Furthermore, seeing the AI generated code first would create "anchoring bias," possibly still resulting in code with more bugs.

                                arnebab@rollenspiel.socialA 1 Reply Last reply
                                0
                                • padjo@mastodon.ieP padjo@mastodon.ie

                                  @Ruth_Mottram @hausfath I had the same experience yesterday. I built a workout tracking app I've been thinking about building for a year. It took about 5 hours of fairly low effort prompting to go from concept to deployed.

                                  Previously this would have been at least a week of full-time high-intensity work. I would probably never would have had the time to do it as a result. These models have fundamentally changed the economics of building software, it's just undeniable at this stage.

                                  arnebab@rollenspiel.socialA This user is from outside of this forum
                                  arnebab@rollenspiel.socialA This user is from outside of this forum
                                  arnebab@rollenspiel.social
                                  wrote last edited by
                                  #46

                                  @Padjo the core question is: for which tasks does this work reliably?

                                  Did you review the code to ensure that it doesn’t have unintended side-effects?

                                  (that’s the difference between having an auto-complete that works on abstract concepts and negligently releasing potentially dangerous products to the public)

                                  ⇒ the fast part is only for the prototyping stage.
                                  @Ruth_Mottram @hausfath

                                  padjo@mastodon.ieP 1 Reply Last reply
                                  0
                                  • 1 1337@techhub.social

                                    @Ruth_Mottram @hausfath This seems like a *really* bad idea. I'm a software engineer and not a scientist, but I believe I've heard there's already a fairly big problem in the sciences with software bugs producing misleading results. I imagine using AI to write code could make this much worse. IMO, the extra time that would've been spent coding everything would not have been wasted. Coding it yourself gives you more time to think about what you're typing and gain a more complete understanding of your code and the libraries you're using; giving you more time and insight to spot bugs or otherwise wrong or less than optimal ways of doing things. If one did a thorough review of the AI generated code to ensure it was correct, I'd guess it take at least the same amount of time. Furthermore, seeing the AI generated code first would create "anchoring bias," possibly still resulting in code with more bugs.

                                    arnebab@rollenspiel.socialA This user is from outside of this forum
                                    arnebab@rollenspiel.socialA This user is from outside of this forum
                                    arnebab@rollenspiel.social
                                    wrote last edited by
                                    #47

                                    @1337 "anchoring bias" is a formulation I searched for.

                                    Thank you!

                                    That anchoring bias is why Larian finally decided not to let their concept artists use AI generated props for inspiration.
                                    @Ruth_Mottram @hausfath

                                    1 Reply Last reply
                                    0
                                    • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                                      I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

                                      Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
                                      Because this kind of performance changes everything and we can't afford to lag behind...
                                      #AIcoding #ML

                                      Edit: here is the climate brink post I mentioned

                                      Link Preview Image
                                      The AI-Augmented Scientist

                                      The promise and pitfalls of using AI tools to boost my capabilities as a scientist

                                      favicon

                                      (www.theclimatebrink.com)

                                      yvandasilva@hachyderm.ioY This user is from outside of this forum
                                      yvandasilva@hachyderm.ioY This user is from outside of this forum
                                      yvandasilva@hachyderm.io
                                      wrote last edited by
                                      #48

                                      @Ruth_Mottram @hausfath it's okay for one shot little scripts.
                                      Which most data science is.

                                      For long term projects that you need to maintain that grow to thousands or millions of lines that need to live long term and be maintained it's not ok.
                                      It adds too much tech debt too quickly.

                                      Writing code was never the problem tbh. Again for scripts and small few pagers, it's as good as any template generator or dumny drag and drop tool.

                                      1 Reply Last reply
                                      0
                                      • pettter@social.accum.seP pettter@social.accum.se

                                        @Ruth_Mottram I have to admit I find his reasoning about energy use misleading at best - he has Claude running for at least around 10 minutes, and is implying that this is comparable in scope to a single ChatGPT query, which is listed as taking 0.3Wh, which is, uh, not comparable. @hausfath

                                        yvandasilva@hachyderm.ioY This user is from outside of this forum
                                        yvandasilva@hachyderm.ioY This user is from outside of this forum
                                        yvandasilva@hachyderm.io
                                        wrote last edited by
                                        #49

                                        @pettter @Ruth_Mottram @hausfath
                                        This is correct, the use of agents which is what allows to have sensible scripts that do what they are supposed to do rather than eyeballing it. Will generate hundreds if not thousands of queries for a very simple input.
                                        Since generally there will be more than one its not unexpected to product multiple thousands of queries via an agent to an LLMs. Its own "thinking mode" and tool triggering will also triggers more queries.
                                        All of that not even going into the "multi-agent" /"swarm of agents" territory.

                                        1 Reply Last reply
                                        0
                                        • arnebab@rollenspiel.socialA arnebab@rollenspiel.social

                                          @benjamingeer As far as I understand it, the task of @Ruth_Mottram was different from the two examples:

                                          - not trying to learn a skill
                                          - not something that’s complex to program, just a time sink (if I understand it correctly)

                                          And there is something in the text by @hausfath that I’ve also seen from others: a management role, detached from development.

                                          Like many scientists who do their data evaluation in Excel or sas GUIs (social sciences). And often don’t understand why it works.
                                          @UlrikeHahn

                                          S This user is from outside of this forum
                                          S This user is from outside of this forum
                                          slotos@toot.community
                                          wrote last edited by
                                          #50

                                          @ArneBab You skipped the most important point:

                                          - not intending for the result to be maintained

                                          For a one-off result these models seem impressive. Hell, outside of a „solve wages” bubble AI field consistently produces useful tools.

                                          But holy shit, can people that have never had to maintain a system after a 10x fuckface has fled the scene shut the fuck up about AI and coding? Code is the easy part where engineers get to finish the productivity reward loop. Go automate your vacations instead!

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups