Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture".

Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture".

Scheduled Pinned Locked Moved Uncategorized
163 Posts 63 Posters 50 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • pluralistic@mamot.frP pluralistic@mamot.fr

    @clintruin @simonzerafa @tante

    Which "couple million people" suffer harm when I run a model on my laptop?

    clintruin@mastodon.socialC This user is from outside of this forum
    clintruin@mastodon.socialC This user is from outside of this forum
    clintruin@mastodon.social
    wrote last edited by
    #128

    @pluralistic @simonzerafa @tante
    Missed the point, sir.

    When one person does it...no big deal.

    When a couple of million people do it...well, see the MIT article above.

    clintruin@mastodon.socialC 1 Reply Last reply
    0
    • pluralistic@mamot.frP pluralistic@mamot.fr

      @tante Dunno where you got the idea that I have a "libertarian" background. I was raised by Trotskyists, am a member of the DSA, am advising and have endorsed Avi Lewis, and joined the UK Greens to back Polanski.

      jorismeys@mstdn.socialJ This user is from outside of this forum
      jorismeys@mstdn.socialJ This user is from outside of this forum
      jorismeys@mstdn.social
      wrote last edited by
      #129

      @pluralistic
      Fair enough, but that's not the core of the argument
      @tante made. He had the same complaint for starters (your argument was heavily drenched in 'you ppl are purists' ), but he also makes the valid argument that technology isn't neutral in itself. Open weights based on intellectual theft and forced labor is still a problem. Until we have a discussion on how the weights come to fruitition, LLM's are objectively problematic from an ethical view. That has nothing to do with purism.

      1 Reply Last reply
      0
      • reflex@retrogaming.socialR This user is from outside of this forum
        reflex@retrogaming.socialR This user is from outside of this forum
        reflex@retrogaming.social
        wrote last edited by
        #130

        @mastodonmigration @shiri @pluralistic @tante The only ethical use of a LLM would be one where the training dataset was ethically acquired, the power was minimized to the level of other methods of providing the same benefits, and the 'benefits' were actually measureable and accurate.

        None of those are true today, and so far as I know there is little to no path to them.

        1 Reply Last reply
        0
        • clintruin@mastodon.socialC clintruin@mastodon.social

          @pluralistic @simonzerafa @tante
          Missed the point, sir.

          When one person does it...no big deal.

          When a couple of million people do it...well, see the MIT article above.

          clintruin@mastodon.socialC This user is from outside of this forum
          clintruin@mastodon.socialC This user is from outside of this forum
          clintruin@mastodon.social
          wrote last edited by
          #131

          @pluralistic @simonzerafa @tante
          Subhead quote from the article:
          "The emissions from individual AI text, image, and video queries seem small—until you add up what the industry isn’t tracking and consider where it’s heading next."

          clintruin@mastodon.socialC pluralistic@mamot.frP 2 Replies Last reply
          0
          • clintruin@mastodon.socialC clintruin@mastodon.social

            @pluralistic @simonzerafa @tante
            Subhead quote from the article:
            "The emissions from individual AI text, image, and video queries seem small—until you add up what the industry isn’t tracking and consider where it’s heading next."

            clintruin@mastodon.socialC This user is from outside of this forum
            clintruin@mastodon.socialC This user is from outside of this forum
            clintruin@mastodon.social
            wrote last edited by
            #132

            @pluralistic @simonzerafa @tante
            But hey, you do you, Cory.
            I'm nobody...your Cory Doctrow.
            Let's all do what Cory does...

            pluralistic@mamot.frP 1 Reply Last reply
            0
            • clintruin@mastodon.socialC clintruin@mastodon.social

              @pluralistic @simonzerafa @tante
              Subhead quote from the article:
              "The emissions from individual AI text, image, and video queries seem small—until you add up what the industry isn’t tracking and consider where it’s heading next."

              pluralistic@mamot.frP This user is from outside of this forum
              pluralistic@mamot.frP This user is from outside of this forum
              pluralistic@mamot.fr
              wrote last edited by
              #133

              @clintruin @simonzerafa @tante

              You are laboring under a misapprehension.

              I will reiterate my question, with all caps for emphasis.

              Which "couple million people" suffer harm when I run a model ON MY LAPTOP?

              clintruin@mastodon.socialC algernon@come-from.mad-scientist.clubA 2 Replies Last reply
              0
              • clintruin@mastodon.socialC clintruin@mastodon.social

                @pluralistic @simonzerafa @tante
                But hey, you do you, Cory.
                I'm nobody...your Cory Doctrow.
                Let's all do what Cory does...

                pluralistic@mamot.frP This user is from outside of this forum
                pluralistic@mamot.frP This user is from outside of this forum
                pluralistic@mamot.fr
                wrote last edited by
                #134

                @clintruin @simonzerafa @tante

                Well, you could "do what Cory does" by familiarizing yourself with the conduct that you are criticizing before engaging in ad hominem.

                To be fair, that's not unique to me, but people who fail to rise to that standard are doing themselves and others no good.

                1 Reply Last reply
                0
                • pluralistic@mamot.frP pluralistic@mamot.fr

                  @clintruin @simonzerafa @tante

                  You are laboring under a misapprehension.

                  I will reiterate my question, with all caps for emphasis.

                  Which "couple million people" suffer harm when I run a model ON MY LAPTOP?

                  clintruin@mastodon.socialC This user is from outside of this forum
                  clintruin@mastodon.socialC This user is from outside of this forum
                  clintruin@mastodon.social
                  wrote last edited by
                  #135

                  @pluralistic @simonzerafa @tante
                  I'll reiterate my response.

                  When you *alone* do it...no big deal.
                  When a couple of million do it ON THEIR OWN LAPTOPS...problem.

                  pluralistic@mamot.frP 1 Reply Last reply
                  0
                  • prinlu@0x.trans.failP prinlu@0x.trans.fail

                    @FediThing @pluralistic @tante i feel in the similar way as big tech has taken the notion of AI and LLMs as a cue/excuse to mount a global campaign of public manipulation and massive investments into a speculative project and pumps gazillions$ into it and convinces everyone it's innevitable tech to be put in bag of potato chips, the backlash is then that anything that bears the name of AI and LLM is poisonous plague and people are unfollowing anyone who's touched it in any way or talks about it in any other way than "it's fascist tech, i'm putting a filter in my feed!" (while it IS fascist tech because it's in hands of fascists).

                    in my view the problem seems not what LLMs are (what kind of tech), but how they are used and what they extract from planet when they are used by the big tech in this monstrous harmful way. of course there's a big blurred line and tech can't be separated from the political, but... AI is not intelligent (Big Tech wants you to believe that), and LLMs are not capable of intelligence and learning (Big Tech wants you to believe that).

                    so i feel like a big chunk of anger and hate should really be directed at techno oligarchs and only partially and much more critically at actual algorithms in play. it's not LLMs that are harming the planet, but rather the extraction, these companies who are absolute evil and are doing whatever the hell they want, unchecked, unregulated.

                    or as varoufakis said to tim nguyen: "we don't want to get rid of your tech or company (google). we want to socialize your company in order to use it more productively" and, if i may add, safely and beneficialy for everyone not just a few.

                    bazkie@beige.partyB This user is from outside of this forum
                    bazkie@beige.partyB This user is from outside of this forum
                    bazkie@beige.party
                    wrote last edited by
                    #136

                    @prinlu @FediThing @pluralistic @tante I agree with most things said in this thread, but on a very practical level, I'm curious what training data was used for the model used by @pluralistic 's typo-checking ollama?

                    for me, that training data is key here. was it consensually allowed for use in training?

                    because as I understand, LLMs need vast amounts of training data, and I'm just not sure how you would get access to such data consensually. would love to be enlightened about this 🙂

                    pluralistic@mamot.frP 1 Reply Last reply
                    0
                    • clintruin@mastodon.socialC clintruin@mastodon.social

                      @pluralistic @simonzerafa @tante
                      I'll reiterate my response.

                      When you *alone* do it...no big deal.
                      When a couple of million do it ON THEIR OWN LAPTOPS...problem.

                      pluralistic@mamot.frP This user is from outside of this forum
                      pluralistic@mamot.frP This user is from outside of this forum
                      pluralistic@mamot.fr
                      wrote last edited by
                      #137

                      @clintruin @simonzerafa @tante

                      OK, sorry, i was under the impression that I was having a discussion with someone who understands this issue.

                      You are completely, empirically, technically wrong.

                      Checking the punctuation on a document on your laptop uses less electricity than watching a Youtube video.

                      clintruin@mastodon.socialC 1 Reply Last reply
                      0
                      • tante@tldr.nettime.orgT tante@tldr.nettime.org

                        Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.

                        Link Preview Image
                        Acting ethically in an imperfect world

                        Life is complicated. Regardless of what your beliefs or politics or ethics are, the way that we set up our society and economy will often force you to act against them: You might not want to fly somewhere but your employer will not accept another mode of transportation, you want to eat vegan but are […]

                        favicon

                        Smashing Frames (tante.cc)

                        johnbrowntypeface@spore.socialJ This user is from outside of this forum
                        johnbrowntypeface@spore.socialJ This user is from outside of this forum
                        johnbrowntypeface@spore.social
                        wrote last edited by
                        #138

                        @tante
                        while we're pointing out logistical inconsistencies..

                        there is zero reason to stop masking in an ongoing pandemic - especially as someone who acknowledged the benefits previously

                        nothing has changed to make this a rational choice and it can't be said to be in solidarity with disabled people (or folks in general)

                        1 Reply Last reply
                        0
                        • tante@tldr.nettime.orgT tante@tldr.nettime.org

                          Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.

                          Link Preview Image
                          Acting ethically in an imperfect world

                          Life is complicated. Regardless of what your beliefs or politics or ethics are, the way that we set up our society and economy will often force you to act against them: You might not want to fly somewhere but your employer will not accept another mode of transportation, you want to eat vegan but are […]

                          favicon

                          Smashing Frames (tante.cc)

                          mallory@hachyderm.ioM This user is from outside of this forum
                          mallory@hachyderm.ioM This user is from outside of this forum
                          mallory@hachyderm.io
                          wrote last edited by
                          #139

                          @tante People like Cory who mock others for their disabilities are not worth paying attention to.

                          1 Reply Last reply
                          0
                          • bazkie@beige.partyB bazkie@beige.party

                            @prinlu @FediThing @pluralistic @tante I agree with most things said in this thread, but on a very practical level, I'm curious what training data was used for the model used by @pluralistic 's typo-checking ollama?

                            for me, that training data is key here. was it consensually allowed for use in training?

                            because as I understand, LLMs need vast amounts of training data, and I'm just not sure how you would get access to such data consensually. would love to be enlightened about this 🙂

                            pluralistic@mamot.frP This user is from outside of this forum
                            pluralistic@mamot.frP This user is from outside of this forum
                            pluralistic@mamot.fr
                            wrote last edited by
                            #140

                            @bazkie @prinlu @FediThing @tante

                            I do not accept the premise that scraping for training data is unethical (leaving aside questions of overloading others' servers).

                            This is how every search engine works. It's how computational linguistics works. It's how the Internet Archive works.

                            Making transient copies of other peoples' work to perform mathematical analysis on them isn't just acceptable, it's an unalloyed good and should be encouraged:

                            Link Preview Image
                            How To Think About Scraping – Pluralistic: Daily links from Cory Doctorow

                            favicon

                            (pluralistic.net)

                            bazkie@beige.partyB 1 Reply Last reply
                            0
                            • shiri@foggyminds.comS This user is from outside of this forum
                              shiri@foggyminds.comS This user is from outside of this forum
                              shiri@foggyminds.com
                              wrote last edited by
                              #141

                              @mastodonmigration
                              it's the "copyright" issue, the outlook that unless everyone who posted anything that was used receives a check for a hefty sum then it's unethical.

                              Copyright is in quotes because it's not really a violation of copyright (the LLMs are not producing whole copies of copywritten materials without basically being forced) nor is it a violation of the intent of copyright (people are confused, copyright was never intended to give artists total control, it's just to ensure new art continues to be created).

                              @pluralistic @reflex @tante

                              reflex@retrogaming.socialR 1 Reply Last reply
                              0
                              • tante@tldr.nettime.orgT tante@tldr.nettime.org

                                Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.

                                Link Preview Image
                                Acting ethically in an imperfect world

                                Life is complicated. Regardless of what your beliefs or politics or ethics are, the way that we set up our society and economy will often force you to act against them: You might not want to fly somewhere but your employer will not accept another mode of transportation, you want to eat vegan but are […]

                                favicon

                                Smashing Frames (tante.cc)

                                n1xnx@tilde.zoneN This user is from outside of this forum
                                n1xnx@tilde.zoneN This user is from outside of this forum
                                n1xnx@tilde.zone
                                wrote last edited by
                                #142

                                @tante
                                I partly agree with Cory and partly not.
                                Refusing to use resource-gobbling datacenter-hosted LLMs makes perfect sense. I'd just as soon heat my house by burning kittens. It is also a rational political statement.

                                Refusing to use an LLM hosted on my own iron is also a political statement, as well as a personal choice. I don't give a hoot about ideological purity; I just distrust clankers, and don't want to get into the habit of depending on them. (Besides, they offer me nothing I cannot as easily do for myself.)

                                1 Reply Last reply
                                0
                                • skyfaller@jawns.clubS skyfaller@jawns.club

                                  @pluralistic I don't think mink fur or LLMs are comparable to criticizing the origins of the internet or transistors. It's the process that produced mink fur and LLMs that is destructive, not merely that it's made by bad people.

                                  For example, LLM crawlers regularly take down independent websites like Codeberg, DDoSing, threatening the small web. You may say "but my LLM is frozen in time, it's not part of that scraping now", but it would not remain useful without updates.

                                  @FediThing @tante

                                  correl@fedi.fenix.lgbtC This user is from outside of this forum
                                  correl@fedi.fenix.lgbtC This user is from outside of this forum
                                  correl@fedi.fenix.lgbt
                                  wrote last edited by
                                  #143

                                  @skyfaller@jawns.club @pluralistic@mamot.fr @FediThing@social.chinwag.org @tante@tldr.nettime.org This is precisely it; it's about the process, not their distance from Altman, Amodei, et al. (which the Ollama project and those like it achieve).

                                  The LLM models themselves are, per this analogy, still almost entirely of the mink-corpse variety, and I think it's a stretch to scream "purity!" at everyone giving you the stink eye for the coat you're wearing.

                                  It's not impossible to have and use a model, locally hosted and energy-efficient, that wasn't directly birthed by mass theft and human abuse (or training directly off of models that were). And having models that aren't, that are genuinely open, is great!
                                  That's how the wickedness gets purged and the underlying tech gets liberated.

                                  Maybe your coat is indeed synthetic, that much is still unclear, because so far all the arguing seems to be focused on the store you got it from and the monsters that operate the worst outlets.

                                  pluralistic@mamot.frP 1 Reply Last reply
                                  0
                                  • pluralistic@mamot.frP pluralistic@mamot.fr

                                    @clintruin @simonzerafa @tante

                                    OK, sorry, i was under the impression that I was having a discussion with someone who understands this issue.

                                    You are completely, empirically, technically wrong.

                                    Checking the punctuation on a document on your laptop uses less electricity than watching a Youtube video.

                                    clintruin@mastodon.socialC This user is from outside of this forum
                                    clintruin@mastodon.socialC This user is from outside of this forum
                                    clintruin@mastodon.social
                                    wrote last edited by
                                    #144

                                    @pluralistic @simonzerafa @tante

                                    Fair enough, Cory. You're gonna do what you want regardless of my accuracy or inaccuracy anyway. And maybe I've misunderstood this. The same way many many will.

                                    But visualize this:

                                    "Hey...I just read Cory Doctrow uses an LLM to check his writing."
                                    "Really?"
                                    "Yeah, it's true."
                                    "Cool, maybe what I've read about ChatGPT is wrong too..."

                                    pluralistic@mamot.frP 1 Reply Last reply
                                    0
                                    • lrhodes@merveilles.townL lrhodes@merveilles.town

                                      I'd actually take this a step further and say that technologies ARE social arrangements.

                                      onepict@chaos.socialO This user is from outside of this forum
                                      onepict@chaos.socialO This user is from outside of this forum
                                      onepict@chaos.social
                                      wrote last edited by
                                      #145

                                      @lrhodes I agree, I believe that we do encode our values into our technology. Particularly with what we code and what we use to code or write.

                                      lrhodes@merveilles.townL 1 Reply Last reply
                                      0
                                      • pluralistic@mamot.frP pluralistic@mamot.fr

                                        @bazkie @prinlu @FediThing @tante

                                        I do not accept the premise that scraping for training data is unethical (leaving aside questions of overloading others' servers).

                                        This is how every search engine works. It's how computational linguistics works. It's how the Internet Archive works.

                                        Making transient copies of other peoples' work to perform mathematical analysis on them isn't just acceptable, it's an unalloyed good and should be encouraged:

                                        Link Preview Image
                                        How To Think About Scraping – Pluralistic: Daily links from Cory Doctorow

                                        favicon

                                        (pluralistic.net)

                                        bazkie@beige.partyB This user is from outside of this forum
                                        bazkie@beige.partyB This user is from outside of this forum
                                        bazkie@beige.party
                                        wrote last edited by
                                        #146

                                        @pluralistic @prinlu @FediThing @tante I think the difference to search engines is how LLM reproduces the training data..

                                        as a thought experiment; what if I'd scrape all your blogposts, then start a blog that makes Cory Doctorow styled blogposts, which would end up more popular than your OG blog since I throw billions in marketing money at it.

                                        would you find that ethical? would you find it acceptable?

                                        further thought experiment; lets say you lose most of your income as a result and have to stop making blogs and start flipping burgers at mcDonalds.

                                        your blog would stop existing, and so, my copycat blog would, too - or at least, it would stop bringing novel blogposts.

                                        this kind of effect is real and will very much hinder cultural development, if not grind it to a halt.

                                        that is a problem - this is culturally unsustainable.

                                        pluralistic@mamot.frP 1 Reply Last reply
                                        0
                                        • clintruin@mastodon.socialC clintruin@mastodon.social

                                          @pluralistic @simonzerafa @tante

                                          Fair enough, Cory. You're gonna do what you want regardless of my accuracy or inaccuracy anyway. And maybe I've misunderstood this. The same way many many will.

                                          But visualize this:

                                          "Hey...I just read Cory Doctrow uses an LLM to check his writing."
                                          "Really?"
                                          "Yeah, it's true."
                                          "Cool, maybe what I've read about ChatGPT is wrong too..."

                                          pluralistic@mamot.frP This user is from outside of this forum
                                          pluralistic@mamot.frP This user is from outside of this forum
                                          pluralistic@mamot.fr
                                          wrote last edited by
                                          #147

                                          @clintruin @simonzerafa @tante

                                          This is an absurd argument.

                                          "I just read about a thing that is fine, but I wasn't paying close attention, so maybe something bad is good?"

                                          Come.

                                          On.

                                          clintruin@mastodon.socialC 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups