Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture".

Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture".

Scheduled Pinned Locked Moved Uncategorized
163 Posts 63 Posters 50 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • skyfaller@jawns.clubS skyfaller@jawns.club

    @pluralistic I don't think mink fur or LLMs are comparable to criticizing the origins of the internet or transistors. It's the process that produced mink fur and LLMs that is destructive, not merely that it's made by bad people.

    For example, LLM crawlers regularly take down independent websites like Codeberg, DDoSing, threatening the small web. You may say "but my LLM is frozen in time, it's not part of that scraping now", but it would not remain useful without updates.

    @FediThing @tante

    correl@fedi.fenix.lgbtC This user is from outside of this forum
    correl@fedi.fenix.lgbtC This user is from outside of this forum
    correl@fedi.fenix.lgbt
    wrote last edited by
    #143

    @skyfaller@jawns.club @pluralistic@mamot.fr @FediThing@social.chinwag.org @tante@tldr.nettime.org This is precisely it; it's about the process, not their distance from Altman, Amodei, et al. (which the Ollama project and those like it achieve).

    The LLM models themselves are, per this analogy, still almost entirely of the mink-corpse variety, and I think it's a stretch to scream "purity!" at everyone giving you the stink eye for the coat you're wearing.

    It's not impossible to have and use a model, locally hosted and energy-efficient, that wasn't directly birthed by mass theft and human abuse (or training directly off of models that were). And having models that aren't, that are genuinely open, is great!
    That's how the wickedness gets purged and the underlying tech gets liberated.

    Maybe your coat is indeed synthetic, that much is still unclear, because so far all the arguing seems to be focused on the store you got it from and the monsters that operate the worst outlets.

    pluralistic@mamot.frP 1 Reply Last reply
    0
    • pluralistic@mamot.frP pluralistic@mamot.fr

      @clintruin @simonzerafa @tante

      OK, sorry, i was under the impression that I was having a discussion with someone who understands this issue.

      You are completely, empirically, technically wrong.

      Checking the punctuation on a document on your laptop uses less electricity than watching a Youtube video.

      clintruin@mastodon.socialC This user is from outside of this forum
      clintruin@mastodon.socialC This user is from outside of this forum
      clintruin@mastodon.social
      wrote last edited by
      #144

      @pluralistic @simonzerafa @tante

      Fair enough, Cory. You're gonna do what you want regardless of my accuracy or inaccuracy anyway. And maybe I've misunderstood this. The same way many many will.

      But visualize this:

      "Hey...I just read Cory Doctrow uses an LLM to check his writing."
      "Really?"
      "Yeah, it's true."
      "Cool, maybe what I've read about ChatGPT is wrong too..."

      pluralistic@mamot.frP 1 Reply Last reply
      0
      • lrhodes@merveilles.townL lrhodes@merveilles.town

        I'd actually take this a step further and say that technologies ARE social arrangements.

        onepict@chaos.socialO This user is from outside of this forum
        onepict@chaos.socialO This user is from outside of this forum
        onepict@chaos.social
        wrote last edited by
        #145

        @lrhodes I agree, I believe that we do encode our values into our technology. Particularly with what we code and what we use to code or write.

        lrhodes@merveilles.townL 1 Reply Last reply
        0
        • pluralistic@mamot.frP pluralistic@mamot.fr

          @bazkie @prinlu @FediThing @tante

          I do not accept the premise that scraping for training data is unethical (leaving aside questions of overloading others' servers).

          This is how every search engine works. It's how computational linguistics works. It's how the Internet Archive works.

          Making transient copies of other peoples' work to perform mathematical analysis on them isn't just acceptable, it's an unalloyed good and should be encouraged:

          Link Preview Image
          How To Think About Scraping – Pluralistic: Daily links from Cory Doctorow

          favicon

          (pluralistic.net)

          bazkie@beige.partyB This user is from outside of this forum
          bazkie@beige.partyB This user is from outside of this forum
          bazkie@beige.party
          wrote last edited by
          #146

          @pluralistic @prinlu @FediThing @tante I think the difference to search engines is how LLM reproduces the training data..

          as a thought experiment; what if I'd scrape all your blogposts, then start a blog that makes Cory Doctorow styled blogposts, which would end up more popular than your OG blog since I throw billions in marketing money at it.

          would you find that ethical? would you find it acceptable?

          further thought experiment; lets say you lose most of your income as a result and have to stop making blogs and start flipping burgers at mcDonalds.

          your blog would stop existing, and so, my copycat blog would, too - or at least, it would stop bringing novel blogposts.

          this kind of effect is real and will very much hinder cultural development, if not grind it to a halt.

          that is a problem - this is culturally unsustainable.

          pluralistic@mamot.frP 1 Reply Last reply
          0
          • clintruin@mastodon.socialC clintruin@mastodon.social

            @pluralistic @simonzerafa @tante

            Fair enough, Cory. You're gonna do what you want regardless of my accuracy or inaccuracy anyway. And maybe I've misunderstood this. The same way many many will.

            But visualize this:

            "Hey...I just read Cory Doctrow uses an LLM to check his writing."
            "Really?"
            "Yeah, it's true."
            "Cool, maybe what I've read about ChatGPT is wrong too..."

            pluralistic@mamot.frP This user is from outside of this forum
            pluralistic@mamot.frP This user is from outside of this forum
            pluralistic@mamot.fr
            wrote last edited by
            #147

            @clintruin @simonzerafa @tante

            This is an absurd argument.

            "I just read about a thing that is fine, but I wasn't paying close attention, so maybe something bad is good?"

            Come.

            On.

            clintruin@mastodon.socialC 1 Reply Last reply
            0
            • jab01701mid@mastodon.socialJ jab01701mid@mastodon.social

              @tante Since I assume all the #Epstein documents have been scraped into all the LLM models by now, I'd love to see an example of LLM tech being used for good.
              Show me the list of Epstein co-conspirators.
              Show me names of who helped them escape accountability, and how they did it.
              Show me who raped children. Their names, addresses, passport photos.
              Then I will believe LLMs and "AI" have delivered a benefit.

              splitmind@rheinneckar.socialS This user is from outside of this forum
              splitmind@rheinneckar.socialS This user is from outside of this forum
              splitmind@rheinneckar.social
              wrote last edited by
              #148

              @jab01701mid Someone made this... https://standardworks.ai/public-archives/epstein-files/documents?q=%28andrew%7E+or+andy%7E%29+AND+NOT+%28%28andy+PRE%2F3+stewart%29+OR+%28andrew%7E+PRE%2F3+plaza%29%29&sort=sexual_content_score%3Adesc%2Cdisturbing_content_score%3Adesc%2Csex_crimes_score%3Adesc

              1 Reply Last reply
              0
              • pluralistic@mamot.frP pluralistic@mamot.fr

                @clintruin @simonzerafa @tante

                This is an absurd argument.

                "I just read about a thing that is fine, but I wasn't paying close attention, so maybe something bad is good?"

                Come.

                On.

                clintruin@mastodon.socialC This user is from outside of this forum
                clintruin@mastodon.socialC This user is from outside of this forum
                clintruin@mastodon.social
                wrote last edited by
                #149

                @pluralistic @simonzerafa @tante
                Maybe...
                Maybe not.

                You have a good day.

                1 Reply Last reply
                0
                • pluralistic@mamot.frP This user is from outside of this forum
                  pluralistic@mamot.frP This user is from outside of this forum
                  pluralistic@mamot.fr
                  wrote last edited by
                  #150

                  @FediThing @bazkie @prinlu @tante

                  There are tons of private search engines, indices, and analysis projects that don't direct text to other works.

                  I could scrape the web for a compilation of "websites no one should visit, ever." That's not "labor theft."

                  1 Reply Last reply
                  0
                  • correl@fedi.fenix.lgbtC correl@fedi.fenix.lgbt

                    @skyfaller@jawns.club @pluralistic@mamot.fr @FediThing@social.chinwag.org @tante@tldr.nettime.org This is precisely it; it's about the process, not their distance from Altman, Amodei, et al. (which the Ollama project and those like it achieve).

                    The LLM models themselves are, per this analogy, still almost entirely of the mink-corpse variety, and I think it's a stretch to scream "purity!" at everyone giving you the stink eye for the coat you're wearing.

                    It's not impossible to have and use a model, locally hosted and energy-efficient, that wasn't directly birthed by mass theft and human abuse (or training directly off of models that were). And having models that aren't, that are genuinely open, is great!
                    That's how the wickedness gets purged and the underlying tech gets liberated.

                    Maybe your coat is indeed synthetic, that much is still unclear, because so far all the arguing seems to be focused on the store you got it from and the monsters that operate the worst outlets.

                    pluralistic@mamot.frP This user is from outside of this forum
                    pluralistic@mamot.frP This user is from outside of this forum
                    pluralistic@mamot.fr
                    wrote last edited by
                    #151

                    @correl @skyfaller @FediThing @tante

                    More fruit of the poisoned tree.

                    "This isn't bad, but it has bad things in its origin. The things I use *also* have bad things in their origin, but that's OK, because those bad things are different because [reasons]."

                    This is the inevitable, pointless dead-end of purity culture.

                    1 Reply Last reply
                    0
                    • onepict@chaos.socialO onepict@chaos.social

                      @lrhodes I agree, I believe that we do encode our values into our technology. Particularly with what we code and what we use to code or write.

                      lrhodes@merveilles.townL This user is from outside of this forum
                      lrhodes@merveilles.townL This user is from outside of this forum
                      lrhodes@merveilles.town
                      wrote last edited by
                      #152

                      @onepict Yeah, code is a pretty literal manifestation of that principle, right?

                      And one of the major advantages of AI from an ideological point of view is that it allows the provider to write their values into *other people's code*.

                      onepict@chaos.socialO 1 Reply Last reply
                      0
                      • pluralistic@mamot.frP This user is from outside of this forum
                        pluralistic@mamot.frP This user is from outside of this forum
                        pluralistic@mamot.fr
                        wrote last edited by
                        #153

                        @taoish @FediThing @tante

                        Because there are no slaves in this instance. Because no one is being harmed or asked to do any work, or being deprived of anything, or adversely affected in *any articulable way*.

                        But yeah, in every other regard, this is exactly that enslaving people.

                        Sure.

                        1 Reply Last reply
                        0
                        • pluralistic@mamot.frP pluralistic@mamot.fr

                          @clintruin @simonzerafa @tante

                          You are laboring under a misapprehension.

                          I will reiterate my question, with all caps for emphasis.

                          Which "couple million people" suffer harm when I run a model ON MY LAPTOP?

                          algernon@come-from.mad-scientist.clubA This user is from outside of this forum
                          algernon@come-from.mad-scientist.clubA This user is from outside of this forum
                          algernon@come-from.mad-scientist.club
                          wrote last edited by
                          #154

                          @pluralistic @clintruin @simonzerafa @tante

                          Which "couple million people" suffer harm when I run a model ON MY LAPTOP?

                          Anyone who's hosting a website, and is getting hammered by the bots that seek content to train the models on. Those of us are the ones who continue getting hurt.

                          Whether you run it locally or not, makes little difference. The models were trained, and training very likely involved scraping, and that continues to be a problem to this day. Not because of ethical concerns, but technical ones: a constant 100req/sec 24/7, with over 2.5k req/sec waves may sound little in this day and age, but at around 2.5k req/sec (sustained for about a week!), my cheap VPS's two vCPUs are bogged down trying to deal with all the TLS handshakes, let alone serving anything.

                          That is a cost many seem to forget. It costs bandwidth, CPU, and human effort to keep things online under the crawler DDoS - which often will require cold, hard cash too, to survive.

                          Ask Codeberg or LWN how they fare under crawler load, and imagine someone who just wants to have their stuff online having to deal with similar abuse.

                          That is the suffering you enable when using any LLM model, even locally.

                          1 Reply Last reply
                          0
                          • lrhodes@merveilles.townL lrhodes@merveilles.town

                            @onepict Yeah, code is a pretty literal manifestation of that principle, right?

                            And one of the major advantages of AI from an ideological point of view is that it allows the provider to write their values into *other people's code*.

                            onepict@chaos.socialO This user is from outside of this forum
                            onepict@chaos.socialO This user is from outside of this forum
                            onepict@chaos.social
                            wrote last edited by
                            #155

                            @lrhodes Yes.

                            There's a desperation as well for some enthusiastic folks to justify this and impose their view on the rest of us. It's what disquieted me. The defensive attitude anticipating us stating and enforcing our boundaries.

                            But it's our culture in Tech and I wish as a whole Tech would step back and like take a minute, rather than reacting and negging.

                            Link Preview Image
                            AI and that Guy at the bar

                            In tech we've always had evangelists, weither it's for FOSS, or Blockchain or now AI. It's a natural thing to do. You have a tech you'r...

                            favicon

                            cobbles (dotart.blog)

                            1 Reply Last reply
                            0
                            • shiri@foggyminds.comS shiri@foggyminds.com

                              @mastodonmigration
                              it's the "copyright" issue, the outlook that unless everyone who posted anything that was used receives a check for a hefty sum then it's unethical.

                              Copyright is in quotes because it's not really a violation of copyright (the LLMs are not producing whole copies of copywritten materials without basically being forced) nor is it a violation of the intent of copyright (people are confused, copyright was never intended to give artists total control, it's just to ensure new art continues to be created).

                              @pluralistic @reflex @tante

                              reflex@retrogaming.socialR This user is from outside of this forum
                              reflex@retrogaming.socialR This user is from outside of this forum
                              reflex@retrogaming.social
                              wrote last edited by
                              #156

                              @shiri @pluralistic @mastodonmigration @tante Also it's incredibly unclear to me how a LLM is a good use case for punctuation and grammar checking,. something regular document editors have done incredibly well since the late 90's or so. Like that's your use case? Not promoting Microsoft here but Word has been fantastic at that since at least 2003.

                              Seems weird to use that as the case for an energy sucking plagiarism machine.

                              1 Reply Last reply
                              0
                              • bazkie@beige.partyB bazkie@beige.party

                                @pluralistic @prinlu @FediThing @tante I think the difference to search engines is how LLM reproduces the training data..

                                as a thought experiment; what if I'd scrape all your blogposts, then start a blog that makes Cory Doctorow styled blogposts, which would end up more popular than your OG blog since I throw billions in marketing money at it.

                                would you find that ethical? would you find it acceptable?

                                further thought experiment; lets say you lose most of your income as a result and have to stop making blogs and start flipping burgers at mcDonalds.

                                your blog would stop existing, and so, my copycat blog would, too - or at least, it would stop bringing novel blogposts.

                                this kind of effect is real and will very much hinder cultural development, if not grind it to a halt.

                                that is a problem - this is culturally unsustainable.

                                pluralistic@mamot.frP This user is from outside of this forum
                                pluralistic@mamot.frP This user is from outside of this forum
                                pluralistic@mamot.fr
                                wrote last edited by
                                #157

                                @bazkie @prinlu @FediThing @tante

                                First: checking for punctuation errors and other typos *in my own work* in a model running on *my own laptop* has nothing - not one single, solitary thing - in common with your example.

                                Nothing.

                                Literally, nothing.

                                But second: I literally license my work for commercial republication and it is widely republished in commercial outlets without any payment or notice to me.

                                bazkie@beige.partyB 1 Reply Last reply
                                0
                                • pluralistic@mamot.frP This user is from outside of this forum
                                  pluralistic@mamot.frP This user is from outside of this forum
                                  pluralistic@mamot.fr
                                  wrote last edited by
                                  #158

                                  @FediThing @bazkie @prinlu @tante

                                  No one is defending "creating knock offs of works." Why would you raise it here? Who has suggested that this is a good way to use LLMs or a good outcome from scraping?

                                  pluralistic@mamot.frP 1 Reply Last reply
                                  0
                                  • pluralistic@mamot.frP pluralistic@mamot.fr

                                    @FediThing @bazkie @prinlu @tante

                                    No one is defending "creating knock offs of works." Why would you raise it here? Who has suggested that this is a good way to use LLMs or a good outcome from scraping?

                                    pluralistic@mamot.frP This user is from outside of this forum
                                    pluralistic@mamot.frP This user is from outside of this forum
                                    pluralistic@mamot.fr
                                    wrote last edited by
                                    #159

                                    @FediThing @bazkie @prinlu @tante

                                    The argument was literally, "It's not OK to check the punctuation in *your own work* if the punctuation checker was created by examining other peoples' work, because performing mathematical analysis on other peoples' work is *per se* unethical."

                                    pluralistic@mamot.frP 1 Reply Last reply
                                    0
                                    • pluralistic@mamot.frP pluralistic@mamot.fr

                                      @FediThing @bazkie @prinlu @tante

                                      The argument was literally, "It's not OK to check the punctuation in *your own work* if the punctuation checker was created by examining other peoples' work, because performing mathematical analysis on other peoples' work is *per se* unethical."

                                      pluralistic@mamot.frP This user is from outside of this forum
                                      pluralistic@mamot.frP This user is from outside of this forum
                                      pluralistic@mamot.fr
                                      wrote last edited by
                                      #160

                                      @FediThing @bazkie @prinlu @tante

                                      By this standard the OED is unethical.

                                      1 Reply Last reply
                                      0
                                      • pluralistic@mamot.frP pluralistic@mamot.fr

                                        @bazkie @prinlu @FediThing @tante

                                        First: checking for punctuation errors and other typos *in my own work* in a model running on *my own laptop* has nothing - not one single, solitary thing - in common with your example.

                                        Nothing.

                                        Literally, nothing.

                                        But second: I literally license my work for commercial republication and it is widely republished in commercial outlets without any payment or notice to me.

                                        bazkie@beige.partyB This user is from outside of this forum
                                        bazkie@beige.partyB This user is from outside of this forum
                                        bazkie@beige.party
                                        wrote last edited by
                                        #161

                                        @pluralistic but then you consented to that, right? you are in control of that.

                                        also my example IS similar - after all, it's data scraped without consent, used to create another work. the typo-checker changes your blogpost based on my training data, in the same way my copycat blog changes 'my' works based on your training data.

                                        sure, it's on a way different scale - deliberately, to more clearly show the principle - but it's the same thing.

                                        pluralistic@mamot.frP 1 Reply Last reply
                                        0
                                        • tante@tldr.nettime.orgT tante@tldr.nettime.org

                                          Yesterday Cory Doctorow argued that refusal to use LLMs was mere "neoliberal purity culture". I think his argument is a strawman, doesn't align with his own actions and delegitimizes important political actions we need to make in order to build a better cyberphysical world.

                                          Link Preview Image
                                          Acting ethically in an imperfect world

                                          Life is complicated. Regardless of what your beliefs or politics or ethics are, the way that we set up our society and economy will often force you to act against them: You might not want to fly somewhere but your employer will not accept another mode of transportation, you want to eat vegan but are […]

                                          favicon

                                          Smashing Frames (tante.cc)

                                          csara@vmst.ioC This user is from outside of this forum
                                          csara@vmst.ioC This user is from outside of this forum
                                          csara@vmst.io
                                          wrote last edited by
                                          #162

                                          @tante I appreciate this post. I have gotten into similar discussions of purity culture around generative AI use (me being against using AI) and you articulate many of the feelings I have about it well.

                                          1 Reply Last reply
                                          1
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups