Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink).

I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink).

Scheduled Pinned Locked Moved Uncategorized
theclimatebrinkaicoding
56 Posts 19 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

    I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

    Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
    Because this kind of performance changes everything and we can't afford to lag behind...
    #AIcoding #ML

    Edit: here is the climate brink post I mentioned

    Link Preview Image
    The AI-Augmented Scientist

    The promise and pitfalls of using AI tools to boost my capabilities as a scientist

    favicon

    (www.theclimatebrink.com)

    wmd@chaos.socialW This user is from outside of this forum
    wmd@chaos.socialW This user is from outside of this forum
    wmd@chaos.social
    wrote last edited by
    #7

    @Ruth_Mottram @hausfath you study climate and want to speed up the change? 🤔 This AI boom might be nice for your personal productivity. But it is also accelerating climate change by it's massive energy consumption.

    1 Reply Last reply
    0
    • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

      I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

      Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
      Because this kind of performance changes everything and we can't afford to lag behind...
      #AIcoding #ML

      Edit: here is the climate brink post I mentioned

      Link Preview Image
      The AI-Augmented Scientist

      The promise and pitfalls of using AI tools to boost my capabilities as a scientist

      favicon

      (www.theclimatebrink.com)

      ulrikehahn@fediscience.orgU This user is from outside of this forum
      ulrikehahn@fediscience.orgU This user is from outside of this forum
      ulrikehahn@fediscience.org
      wrote last edited by
      #8

      @Ruth_Mottram @hausfath are you aware of this project: https://www.swiss-ai.org/apertus

      ulrikehahn@fediscience.orgU 1 Reply Last reply
      0
      • normis@g.dodies.lvN normis@g.dodies.lv

        @Ruth_Mottram @hausfath once you try the claude code terminal, ask it to use planning mode and agent teams, it is amazing. The web chat bot does not compare. For european alternatives look into "Mistral Vibe"

        ruth_mottram@fediscience.orgR This user is from outside of this forum
        ruth_mottram@fediscience.orgR This user is from outside of this forum
        ruth_mottram@fediscience.org
        wrote last edited by
        #9

        @normis @hausfath thanks!

        1 Reply Last reply
        0
        • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

          @pettter @hausfath it took me half an hour to do something that would normally have taken at least a day (and for which reason I have been procrastinating over it)

          This piece on the climate brink is spot on:

          Link Preview Image
          The AI-Augmented Scientist

          The promise and pitfalls of using AI tools to boost my capabilities as a scientist

          favicon

          (www.theclimatebrink.com)

          pettter@social.accum.seP This user is from outside of this forum
          pettter@social.accum.seP This user is from outside of this forum
          pettter@social.accum.se
          wrote last edited by
          #10

          @Ruth_Mottram So if I understand this correctly, this is about not having to read documentation or know the software tools one is using? @hausfath

          pettter@social.accum.seP 1 Reply Last reply
          0
          • pettter@social.accum.seP pettter@social.accum.se

            @Ruth_Mottram So if I understand this correctly, this is about not having to read documentation or know the software tools one is using? @hausfath

            pettter@social.accum.seP This user is from outside of this forum
            pettter@social.accum.seP This user is from outside of this forum
            pettter@social.accum.se
            wrote last edited by
            #11

            @Ruth_Mottram I have to admit I find his reasoning about energy use misleading at best - he has Claude running for at least around 10 minutes, and is implying that this is comparable in scope to a single ChatGPT query, which is listed as taking 0.3Wh, which is, uh, not comparable. @hausfath

            yvandasilva@hachyderm.ioY 1 Reply Last reply
            0
            • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

              @Ruth_Mottram @hausfath are you aware of this project: https://www.swiss-ai.org/apertus

              ulrikehahn@fediscience.orgU This user is from outside of this forum
              ulrikehahn@fediscience.orgU This user is from outside of this forum
              ulrikehahn@fediscience.org
              wrote last edited by
              #12

              @Ruth_Mottram @hausfath and if these systems were “just hype” we wouldn’t be seeing the widespread systemic threats and transformations we are now in the midst of - the hype discourse did nothing to prepare us for the damage and did us a disservice if you ask me

              ruth_mottram@fediscience.orgR 1 Reply Last reply
              0
              • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                @Ruth_Mottram @hausfath and if these systems were “just hype” we wouldn’t be seeing the widespread systemic threats and transformations we are now in the midst of - the hype discourse did nothing to prepare us for the damage and did us a disservice if you ask me

                ruth_mottram@fediscience.orgR This user is from outside of this forum
                ruth_mottram@fediscience.orgR This user is from outside of this forum
                ruth_mottram@fediscience.org
                wrote last edited by
                #13

                @UlrikeHahn @hausfath yes I think so too!

                benjamingeer@piaille.frB 1 Reply Last reply
                0
                • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                  @UlrikeHahn @hausfath yes I think so too!

                  benjamingeer@piaille.frB This user is from outside of this forum
                  benjamingeer@piaille.frB This user is from outside of this forum
                  benjamingeer@piaille.fr
                  wrote last edited by
                  #14

                  @Ruth_Mottram @UlrikeHahn @hausfath But so far nobody has found evidence of productivity gains in a controlled experiment, not even Anthropic? https://www.anthropic.com/research/AI-assistance-coding-skills One experiment found that LLM coding assistants made developers less productive even though the developers believed it made them more productive. https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/

                  ulrikehahn@fediscience.orgU arnebab@rollenspiel.socialA 2 Replies Last reply
                  0
                  • benjamingeer@piaille.frB benjamingeer@piaille.fr

                    @Ruth_Mottram @UlrikeHahn @hausfath But so far nobody has found evidence of productivity gains in a controlled experiment, not even Anthropic? https://www.anthropic.com/research/AI-assistance-coding-skills One experiment found that LLM coding assistants made developers less productive even though the developers believed it made them more productive. https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/

                    ulrikehahn@fediscience.orgU This user is from outside of this forum
                    ulrikehahn@fediscience.orgU This user is from outside of this forum
                    ulrikehahn@fediscience.org
                    wrote last edited by
                    #15

                    @benjamingeer @Ruth_Mottram @hausfath there are massive “productivity gains” everywhere:

                    - higher education is demonstrably being undermined by the fact that course work can now be (and is being) completed by AI query
                    - science is demonstrably buckling under a deluge of submissions
                    - democracy is demonstrably being harmed by AI based astroturfing

                    all of these are well-documented. “Controlled experiments” are not the only form of evidence ….

                    ruth_mottram@fediscience.orgR benjamingeer@piaille.frB garonenur@rollenspiel.socialG 3 Replies Last reply
                    0
                    • leonoverweel@mastodon.socialL leonoverweel@mastodon.social

                      @Ruth_Mottram by now it’s just a loud minority on here that’s doing the flaming, at least looking at this fairly large poll from last week: https://mastodon.social/@stroughtonsmith/116057255985985154

                      krnlg@mastodon.socialK This user is from outside of this forum
                      krnlg@mastodon.socialK This user is from outside of this forum
                      krnlg@mastodon.social
                      wrote last edited by
                      #16

                      @leonoverweel
                      Depends entirely on who follows the poll author. I've seen polls 97% against AI. The scientific approach would be to study what the productivity effects actually are over time. Personally I don't think it is worth the dependence and lack of understanding, exploitation of programmer's labour and so forth even if it helps someone be more productive. We're already all 1000s of times more productive than pre-computer days and people still die homeless.
                      @Ruth_Mottram

                      1 Reply Last reply
                      0
                      • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                        @benjamingeer @Ruth_Mottram @hausfath there are massive “productivity gains” everywhere:

                        - higher education is demonstrably being undermined by the fact that course work can now be (and is being) completed by AI query
                        - science is demonstrably buckling under a deluge of submissions
                        - democracy is demonstrably being harmed by AI based astroturfing

                        all of these are well-documented. “Controlled experiments” are not the only form of evidence ….

                        ruth_mottram@fediscience.orgR This user is from outside of this forum
                        ruth_mottram@fediscience.orgR This user is from outside of this forum
                        ruth_mottram@fediscience.org
                        wrote last edited by
                        #17

                        @UlrikeHahn @benjamingeer @hausfath yes indeed - I think this piece in the FT, part of @jbu's series is good on this too. It's definitely my experience working with young students - They're much faster getting code up and running than they used to be
                        Is AI making work more intense? - https://giftarticle.ft.com/giftarticle/actions/redeem/eaf0cedf-d53a-4053-9d83-f8b62d39316a via @FT

                        1 Reply Last reply
                        0
                        • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                          I realise on the fediverse this is maybe asking for a flaming, but yesterday out of sheer curiosity I tried Claude for a simpleish coding task that I'd been putting off (largely inspired by @hausfath 's latest on #theclimatebrink). The performance of Claude was seriously impressive. I am convinced the AI cycle is more than hype (and have been for a while), the chatbots have been a huge attention hogger, misleadingly so, while the serious work has been done elsewhere. (We are developing ML tools to supplement parts of our climate model workflows).

                          Now I'm wondering if there is any serious EU competition to Anthropic? - Mistral's codestral perhaps?
                          Because this kind of performance changes everything and we can't afford to lag behind...
                          #AIcoding #ML

                          Edit: here is the climate brink post I mentioned

                          Link Preview Image
                          The AI-Augmented Scientist

                          The promise and pitfalls of using AI tools to boost my capabilities as a scientist

                          favicon

                          (www.theclimatebrink.com)

                          padjo@mastodon.ieP This user is from outside of this forum
                          padjo@mastodon.ieP This user is from outside of this forum
                          padjo@mastodon.ie
                          wrote last edited by
                          #18

                          @Ruth_Mottram @hausfath I had the same experience yesterday. I built a workout tracking app I've been thinking about building for a year. It took about 5 hours of fairly low effort prompting to go from concept to deployed.

                          Previously this would have been at least a week of full-time high-intensity work. I would probably never would have had the time to do it as a result. These models have fundamentally changed the economics of building software, it's just undeniable at this stage.

                          arnebab@rollenspiel.socialA 1 Reply Last reply
                          0
                          • benjamingeer@piaille.frB benjamingeer@piaille.fr

                            @Ruth_Mottram @UlrikeHahn @hausfath But so far nobody has found evidence of productivity gains in a controlled experiment, not even Anthropic? https://www.anthropic.com/research/AI-assistance-coding-skills One experiment found that LLM coding assistants made developers less productive even though the developers believed it made them more productive. https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/

                            arnebab@rollenspiel.socialA This user is from outside of this forum
                            arnebab@rollenspiel.socialA This user is from outside of this forum
                            arnebab@rollenspiel.social
                            wrote last edited by
                            #19

                            @benjamingeer As far as I understand it, the task of @Ruth_Mottram was different from the two examples:

                            - not trying to learn a skill
                            - not something that’s complex to program, just a time sink (if I understand it correctly)

                            And there is something in the text by @hausfath that I’ve also seen from others: a management role, detached from development.

                            Like many scientists who do their data evaluation in Excel or sas GUIs (social sciences). And often don’t understand why it works.
                            @UlrikeHahn

                            ruth_mottram@fediscience.orgR benjamingeer@piaille.frB S 3 Replies Last reply
                            0
                            • ulrikehahn@fediscience.orgU ulrikehahn@fediscience.org

                              @benjamingeer @Ruth_Mottram @hausfath there are massive “productivity gains” everywhere:

                              - higher education is demonstrably being undermined by the fact that course work can now be (and is being) completed by AI query
                              - science is demonstrably buckling under a deluge of submissions
                              - democracy is demonstrably being harmed by AI based astroturfing

                              all of these are well-documented. “Controlled experiments” are not the only form of evidence ….

                              benjamingeer@piaille.frB This user is from outside of this forum
                              benjamingeer@piaille.frB This user is from outside of this forum
                              benjamingeer@piaille.fr
                              wrote last edited by
                              #20

                              @UlrikeHahn Is slop productivity? LLMs are good at producing fake course work, fake scientific papers, fake political debates, etc., which can look plausible and often pass for the real thing if you don't look too closely. @Ruth_Mottram @hausfath

                              ulrikehahn@fediscience.orgU 1 Reply Last reply
                              0
                              • arnebab@rollenspiel.socialA arnebab@rollenspiel.social

                                @benjamingeer As far as I understand it, the task of @Ruth_Mottram was different from the two examples:

                                - not trying to learn a skill
                                - not something that’s complex to program, just a time sink (if I understand it correctly)

                                And there is something in the text by @hausfath that I’ve also seen from others: a management role, detached from development.

                                Like many scientists who do their data evaluation in Excel or sas GUIs (social sciences). And often don’t understand why it works.
                                @UlrikeHahn

                                ruth_mottram@fediscience.orgR This user is from outside of this forum
                                ruth_mottram@fediscience.orgR This user is from outside of this forum
                                ruth_mottram@fediscience.org
                                wrote last edited by
                                #21

                                @ArneBab @benjamingeer @hausfath @UlrikeHahn yes that's exactly the kind of task I think the ml models work well on. A lot of science is actually quite boring and repetitive but needs careful monitoring. If a tool can do part of that. Then why not. I think Zeke is correct in that the human mind needs to come up with the creativity and the experiments as well as with careful analysis to understand the results.

                                arnebab@rollenspiel.socialA benjamingeer@piaille.frB 2 Replies Last reply
                                0
                                • arnebab@rollenspiel.socialA arnebab@rollenspiel.social

                                  @benjamingeer As far as I understand it, the task of @Ruth_Mottram was different from the two examples:

                                  - not trying to learn a skill
                                  - not something that’s complex to program, just a time sink (if I understand it correctly)

                                  And there is something in the text by @hausfath that I’ve also seen from others: a management role, detached from development.

                                  Like many scientists who do their data evaluation in Excel or sas GUIs (social sciences). And often don’t understand why it works.
                                  @UlrikeHahn

                                  benjamingeer@piaille.frB This user is from outside of this forum
                                  benjamingeer@piaille.frB This user is from outside of this forum
                                  benjamingeer@piaille.fr
                                  wrote last edited by
                                  #22

                                  @ArneBab It's true that scientists use calculators even though many of them probably don't really know how calculators work. But if you bought a calculator that sometimes said 2 + 2 = 5, you'd return it and get a refund. LLMs are like that.

                                  LLMs can certainly generate a lot of code very fast. But is it good code, or a mass of spaghetti? Will you be able to maintain it, considering that you don't know how it works? When it turns out to have bugs, will you be able to fix them?

                                  @Ruth_Mottram @hausfath @UlrikeHahn

                                  arnebab@rollenspiel.socialA 1 Reply Last reply
                                  0
                                  • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                                    @ArneBab @benjamingeer @hausfath @UlrikeHahn yes that's exactly the kind of task I think the ml models work well on. A lot of science is actually quite boring and repetitive but needs careful monitoring. If a tool can do part of that. Then why not. I think Zeke is correct in that the human mind needs to come up with the creativity and the experiments as well as with careful analysis to understand the results.

                                    arnebab@rollenspiel.socialA This user is from outside of this forum
                                    arnebab@rollenspiel.socialA This user is from outside of this forum
                                    arnebab@rollenspiel.social
                                    wrote last edited by
                                    #23

                                    @Ruth_Mottram The main risk I see with that is that it can quickly limit creativity.

                                    I experimented with ChatGPT for writing (but didn’t make the results public, except for an experiment explicitly done for evaluation of its effects -- worrying¹), and I found that it is good at providing a start, but repetitive, so that when I started with it, it limited imagination -- kind of like an effect of advertisements. So it’s a bad start.

                                    ¹ https://www.draketo.de/software/ai-translation-evaluated#completely-changed
                                    @benjamingeer @hausfath @UlrikeHahn

                                    arnebab@rollenspiel.socialA 1 Reply Last reply
                                    0
                                    • benjamingeer@piaille.frB benjamingeer@piaille.fr

                                      @UlrikeHahn Is slop productivity? LLMs are good at producing fake course work, fake scientific papers, fake political debates, etc., which can look plausible and often pass for the real thing if you don't look too closely. @Ruth_Mottram @hausfath

                                      ulrikehahn@fediscience.orgU This user is from outside of this forum
                                      ulrikehahn@fediscience.orgU This user is from outside of this forum
                                      ulrikehahn@fediscience.org
                                      wrote last edited by
                                      #24

                                      @benjamingeer @Ruth_Mottram @hausfath I take “productivity” to refer to the efficiency of production of a good or service

                                      readily available AI systems now can (and do) produce essay answers that I would have to assign a passing grade (and actually an increasingly good grade) given our marking criteria, and it can do that in seconds. It’s a huge problem for higher education.

                                      How is that not a “productivity gain”?

                                      I find the conflation of questions about what these systems can actually do (an empirical question!) with questions of desirability deeply counter-productive.

                                      benjamingeer@piaille.frB 1 Reply Last reply
                                      0
                                      • ruth_mottram@fediscience.orgR ruth_mottram@fediscience.org

                                        @ArneBab @benjamingeer @hausfath @UlrikeHahn yes that's exactly the kind of task I think the ml models work well on. A lot of science is actually quite boring and repetitive but needs careful monitoring. If a tool can do part of that. Then why not. I think Zeke is correct in that the human mind needs to come up with the creativity and the experiments as well as with careful analysis to understand the results.

                                        benjamingeer@piaille.frB This user is from outside of this forum
                                        benjamingeer@piaille.frB This user is from outside of this forum
                                        benjamingeer@piaille.fr
                                        wrote last edited by
                                        #25

                                        @Ruth_Mottram The risk is in the "careful monitoring" part: https://mastodon.online/@pseudonym/116135917950981989 @ArneBab @hausfath @UlrikeHahn

                                        1 Reply Last reply
                                        0
                                        • arnebab@rollenspiel.socialA arnebab@rollenspiel.social

                                          @Ruth_Mottram The main risk I see with that is that it can quickly limit creativity.

                                          I experimented with ChatGPT for writing (but didn’t make the results public, except for an experiment explicitly done for evaluation of its effects -- worrying¹), and I found that it is good at providing a start, but repetitive, so that when I started with it, it limited imagination -- kind of like an effect of advertisements. So it’s a bad start.

                                          ¹ https://www.draketo.de/software/ai-translation-evaluated#completely-changed
                                          @benjamingeer @hausfath @UlrikeHahn

                                          arnebab@rollenspiel.socialA This user is from outside of this forum
                                          arnebab@rollenspiel.socialA This user is from outside of this forum
                                          arnebab@rollenspiel.social
                                          wrote last edited by
                                          #26

                                          @Ruth_Mottram One experiment I did was to turn a text I wrote years ago into a scientific paper in economics.

                                          It took two hours and reached a quality that I (physicist, not from economics) could not have distinguished it from a real paper.

                                          AI causes the form to be easier to repeat, so we can no longer trust the form of scientific writing to be a hint that people actually have scientific education.

                                          And that is a huge risk.
                                          @benjamingeer @hausfath @UlrikeHahn

                                          arnebab@rollenspiel.socialA 2 Replies Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups