Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Yesterday, had an argument with an AI booster.

Yesterday, had an argument with an AI booster.

Scheduled Pinned Locked Moved Uncategorized
31 Posts 23 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • xgranade@wandering.shopX xgranade@wandering.shop

    Yesterday, had an argument with an AI booster. I'm not going to link, both because I don't want to platform that and because I don't want anyone to go harass them. But what I thought was very interesting was that I asked point-blank if there was any degree to which ethical problems with LLMs could make them not want to use AI — they told me no, there was not, and implied that they evaluated AI purely on the basis of its efficacy.

    xgranade@wandering.shopX This user is from outside of this forum
    xgranade@wandering.shopX This user is from outside of this forum
    xgranade@wandering.shop
    wrote last edited by
    #2

    I don't have time nor the inclination to argue that point with them further when it comes to AI. But I do think there's a broader point that is worth critical examination, especially as tech continues to build out surveillance, age verification, automated filtering and censoring, and other tools that do immense damage when used by authoritarians.

    We *cannot* afford to evaluate tech purely based on whether it "works" or not.

    xgranade@wandering.shopX tankgrrl@hachyderm.ioT 2 Replies Last reply
    0
    • xgranade@wandering.shopX xgranade@wandering.shop

      I don't have time nor the inclination to argue that point with them further when it comes to AI. But I do think there's a broader point that is worth critical examination, especially as tech continues to build out surveillance, age verification, automated filtering and censoring, and other tools that do immense damage when used by authoritarians.

      We *cannot* afford to evaluate tech purely based on whether it "works" or not.

      xgranade@wandering.shopX This user is from outside of this forum
      xgranade@wandering.shopX This user is from outside of this forum
      xgranade@wandering.shop
      wrote last edited by
      #3

      AI doesn't work¹, so it's easy to forget that larger point, I suspect? That *even if* AI did work (and again, it doesn't), it still would need to be critically examined from an ethical perspective.

      Failing to do so is how we have massive surveillance networks today.

      ___
      ¹Here again, referring to the wave of current hype products. Boosters love wearing the ML shit that does work as a shield against criticism.

      yosh@toot.yosh.isY kevingranade@mastodon.gamedev.placeK pythonbynight@hachyderm.ioP newbery@mastodon.socialN tuban_muzuru@ohai.socialT 7 Replies Last reply
      1
      0
      • xgranade@wandering.shopX xgranade@wandering.shop

        AI doesn't work¹, so it's easy to forget that larger point, I suspect? That *even if* AI did work (and again, it doesn't), it still would need to be critically examined from an ethical perspective.

        Failing to do so is how we have massive surveillance networks today.

        ___
        ¹Here again, referring to the wave of current hype products. Boosters love wearing the ML shit that does work as a shield against criticism.

        yosh@toot.yosh.isY This user is from outside of this forum
        yosh@toot.yosh.isY This user is from outside of this forum
        yosh@toot.yosh.is
        wrote last edited by
        #4

        @xgranade

        I'm not sure I'd say "AI doesn't work" anymore. It definitely doesn't "work" to the degree that the loudest boosters will claim it does. But like, I do think it's recently crossed a threshold where it can be a useful tool in the right hands.

        Which I personally find very annoying since I too have moral qualms about the broader AI industry. E.g. the point about surveillance you're making I think is an important one.

        mms@mastodon.bsd.cafeM jamwil@mastodon.sdf.orgJ 2 Replies Last reply
        0
        • xgranade@wandering.shopX xgranade@wandering.shop

          Yesterday, had an argument with an AI booster. I'm not going to link, both because I don't want to platform that and because I don't want anyone to go harass them. But what I thought was very interesting was that I asked point-blank if there was any degree to which ethical problems with LLMs could make them not want to use AI — they told me no, there was not, and implied that they evaluated AI purely on the basis of its efficacy.

          bstacey@icosahedron.websiteB This user is from outside of this forum
          bstacey@icosahedron.websiteB This user is from outside of this forum
          bstacey@icosahedron.website
          wrote last edited by
          #5

          @xgranade jfc

          1 Reply Last reply
          0
          • xgranade@wandering.shopX xgranade@wandering.shop

            AI doesn't work¹, so it's easy to forget that larger point, I suspect? That *even if* AI did work (and again, it doesn't), it still would need to be critically examined from an ethical perspective.

            Failing to do so is how we have massive surveillance networks today.

            ___
            ¹Here again, referring to the wave of current hype products. Boosters love wearing the ML shit that does work as a shield against criticism.

            kevingranade@mastodon.gamedev.placeK This user is from outside of this forum
            kevingranade@mastodon.gamedev.placeK This user is from outside of this forum
            kevingranade@mastodon.gamedev.place
            wrote last edited by
            #6

            @xgranade it's exhausting as fuck when people that both agree with and disagree with your main point (don't use or promote LLMs) object when the rationales you muster in support are not precisely the same ones they care about.
            IMO this is where people get the idea that the "purity culture" angle presents a wedge issue they can use to attack anti-AI people.

            xgranade@wandering.shopX 1 Reply Last reply
            0
            • kevingranade@mastodon.gamedev.placeK kevingranade@mastodon.gamedev.place

              @xgranade it's exhausting as fuck when people that both agree with and disagree with your main point (don't use or promote LLMs) object when the rationales you muster in support are not precisely the same ones they care about.
              IMO this is where people get the idea that the "purity culture" angle presents a wedge issue they can use to attack anti-AI people.

              xgranade@wandering.shopX This user is from outside of this forum
              xgranade@wandering.shopX This user is from outside of this forum
              xgranade@wandering.shop
              wrote last edited by
              #7

              @kevingranade I never want to put my AI Luddism on a pedestal and make it immune to critique... this is, for example, why I said my closed-mindedness on the subject is both temporary and a reasoned response to bad-faith DDoS attacks on discourse.

              To that extent, I'm glad for critique from "my" side. But the purity culture discourse (with a few important exceptions) isn't that, it's a wedge.

              xgranade@wandering.shopX 1 Reply Last reply
              0
              • xgranade@wandering.shopX xgranade@wandering.shop

                @kevingranade I never want to put my AI Luddism on a pedestal and make it immune to critique... this is, for example, why I said my closed-mindedness on the subject is both temporary and a reasoned response to bad-faith DDoS attacks on discourse.

                To that extent, I'm glad for critique from "my" side. But the purity culture discourse (with a few important exceptions) isn't that, it's a wedge.

                xgranade@wandering.shopX This user is from outside of this forum
                xgranade@wandering.shopX This user is from outside of this forum
                xgranade@wandering.shop
                wrote last edited by
                #8

                @kevingranade I find that progressive moments, as a consequence of the laudable and correct willingness to self-criticize, tend to be vulnerable to wedge attacks. Fuck, as a trans person I *am* a wedge, or at least the right-wing has turned me into a wedge used to weaken opposition to violent and cruel immigration policies.

                We need to get a lot better at distinguishing wedges from critiques.

                1 Reply Last reply
                0
                • xgranade@wandering.shopX xgranade@wandering.shop

                  Yesterday, had an argument with an AI booster. I'm not going to link, both because I don't want to platform that and because I don't want anyone to go harass them. But what I thought was very interesting was that I asked point-blank if there was any degree to which ethical problems with LLMs could make them not want to use AI — they told me no, there was not, and implied that they evaluated AI purely on the basis of its efficacy.

                  wbftw@hachyderm.ioW This user is from outside of this forum
                  wbftw@hachyderm.ioW This user is from outside of this forum
                  wbftw@hachyderm.io
                  wrote last edited by
                  #9

                  @xgranade had a very similar conversation the other day too; only was able to somewhat shift my interlocutor’s position after pointing out they don’t own this “tool”, and they are at the mercy of fash/oligarch class who can (and will) start extracting rent at any moment.

                  riotnrrd@mastodon.socialR 1 Reply Last reply
                  0
                  • xgranade@wandering.shopX xgranade@wandering.shop

                    Yesterday, had an argument with an AI booster. I'm not going to link, both because I don't want to platform that and because I don't want anyone to go harass them. But what I thought was very interesting was that I asked point-blank if there was any degree to which ethical problems with LLMs could make them not want to use AI — they told me no, there was not, and implied that they evaluated AI purely on the basis of its efficacy.

                    leastaction@writing.exchangeL This user is from outside of this forum
                    leastaction@writing.exchangeL This user is from outside of this forum
                    leastaction@writing.exchange
                    wrote last edited by
                    #10

                    @xgranade Its the government's job to regulate AI on an ethical basis, on behalf of all of us, because individually we don't have the power to do that, and the government represents the people in a democracy, and it does have the power. The fact that the government is not doing this at all shows to what extent we do not have a democracy. Not even close.

                    neongod@mstdn.socialN 1 Reply Last reply
                    0
                    • xgranade@wandering.shopX xgranade@wandering.shop

                      AI doesn't work¹, so it's easy to forget that larger point, I suspect? That *even if* AI did work (and again, it doesn't), it still would need to be critically examined from an ethical perspective.

                      Failing to do so is how we have massive surveillance networks today.

                      ___
                      ¹Here again, referring to the wave of current hype products. Boosters love wearing the ML shit that does work as a shield against criticism.

                      pythonbynight@hachyderm.ioP This user is from outside of this forum
                      pythonbynight@hachyderm.ioP This user is from outside of this forum
                      pythonbynight@hachyderm.io
                      wrote last edited by
                      #11

                      @xgranade This has also been fascinating to me lately.

                      My latest blog post is specifically musing on what that calculus looks like for each person.

                      I always presumed that a major factor is in the belief that harms are not as bad as reported (which calls into question what sources we are viewing as authoritative) or that the current and/or future benefits to humanity are worth it (current harms are collateral damage in lieu of progress).

                      But for a person to be unable to imagine any scenario that would change their mind seems crazy to me.

                      I often ask myself what it would take to change my mind on this issue, and while I think most of those scenarios are highly improbable.. I can still imagine them.

                      pythonbynight@hachyderm.ioP 1 Reply Last reply
                      0
                      • pythonbynight@hachyderm.ioP pythonbynight@hachyderm.io

                        @xgranade This has also been fascinating to me lately.

                        My latest blog post is specifically musing on what that calculus looks like for each person.

                        I always presumed that a major factor is in the belief that harms are not as bad as reported (which calls into question what sources we are viewing as authoritative) or that the current and/or future benefits to humanity are worth it (current harms are collateral damage in lieu of progress).

                        But for a person to be unable to imagine any scenario that would change their mind seems crazy to me.

                        I often ask myself what it would take to change my mind on this issue, and while I think most of those scenarios are highly improbable.. I can still imagine them.

                        pythonbynight@hachyderm.ioP This user is from outside of this forum
                        pythonbynight@hachyderm.ioP This user is from outside of this forum
                        pythonbynight@hachyderm.io
                        wrote last edited by
                        #12

                        @xgranade hmm, now that I think about it, maybe I should articulate that in a future blog post.

                        1 Reply Last reply
                        0
                        • xgranade@wandering.shopX xgranade@wandering.shop

                          AI doesn't work¹, so it's easy to forget that larger point, I suspect? That *even if* AI did work (and again, it doesn't), it still would need to be critically examined from an ethical perspective.

                          Failing to do so is how we have massive surveillance networks today.

                          ___
                          ¹Here again, referring to the wave of current hype products. Boosters love wearing the ML shit that does work as a shield against criticism.

                          newbery@mastodon.socialN This user is from outside of this forum
                          newbery@mastodon.socialN This user is from outside of this forum
                          newbery@mastodon.social
                          wrote last edited by
                          #13

                          @xgranade
                          In a meeting today, I recently raised the analogy to debates about torture. The debate gets muddy when people mix together questions of efficacy with questions of ethics and morality. Sometimes this confusion is deliberate but I suspect most of the time, it's just a lack of clarity on priorities. Whatever action we are considering, the first question should always be whether it's an ethical or moral action. The efficacy question should always be secondary.

                          1 Reply Last reply
                          0
                          • xgranade@wandering.shopX xgranade@wandering.shop

                            Yesterday, had an argument with an AI booster. I'm not going to link, both because I don't want to platform that and because I don't want anyone to go harass them. But what I thought was very interesting was that I asked point-blank if there was any degree to which ethical problems with LLMs could make them not want to use AI — they told me no, there was not, and implied that they evaluated AI purely on the basis of its efficacy.

                            tankgrrl@hachyderm.ioT This user is from outside of this forum
                            tankgrrl@hachyderm.ioT This user is from outside of this forum
                            tankgrrl@hachyderm.io
                            wrote last edited by
                            #14

                            @xgranade
                            Sociopaths gonna socio... path.

                            1 Reply Last reply
                            0
                            • xgranade@wandering.shopX xgranade@wandering.shop

                              I don't have time nor the inclination to argue that point with them further when it comes to AI. But I do think there's a broader point that is worth critical examination, especially as tech continues to build out surveillance, age verification, automated filtering and censoring, and other tools that do immense damage when used by authoritarians.

                              We *cannot* afford to evaluate tech purely based on whether it "works" or not.

                              tankgrrl@hachyderm.ioT This user is from outside of this forum
                              tankgrrl@hachyderm.ioT This user is from outside of this forum
                              tankgrrl@hachyderm.io
                              wrote last edited by
                              #15

                              @xgranade And especially given that it works reliably 'in this niche, but not this one'. Reliability and repeatability is all over the map.

                              1 Reply Last reply
                              0
                              • xgranade@wandering.shopX xgranade@wandering.shop

                                Yesterday, had an argument with an AI booster. I'm not going to link, both because I don't want to platform that and because I don't want anyone to go harass them. But what I thought was very interesting was that I asked point-blank if there was any degree to which ethical problems with LLMs could make them not want to use AI — they told me no, there was not, and implied that they evaluated AI purely on the basis of its efficacy.

                                tattooed_mummy@beige.partyT This user is from outside of this forum
                                tattooed_mummy@beige.partyT This user is from outside of this forum
                                tattooed_mummy@beige.party
                                wrote last edited by
                                #16

                                @xgranade wow

                                1 Reply Last reply
                                0
                                • leastaction@writing.exchangeL leastaction@writing.exchange

                                  @xgranade Its the government's job to regulate AI on an ethical basis, on behalf of all of us, because individually we don't have the power to do that, and the government represents the people in a democracy, and it does have the power. The fact that the government is not doing this at all shows to what extent we do not have a democracy. Not even close.

                                  neongod@mstdn.socialN This user is from outside of this forum
                                  neongod@mstdn.socialN This user is from outside of this forum
                                  neongod@mstdn.social
                                  wrote last edited by
                                  #17

                                  @leastaction @xgranade it just shows that most people (represented by politicians) don’t care. They care much more about economic growth and their wealth then ethics. Also people in general are extremely lazy, which is why it is so tempting to use it and most are even willing to outsource their thinking to it.

                                  neongod@mstdn.socialN 1 Reply Last reply
                                  0
                                  • neongod@mstdn.socialN neongod@mstdn.social

                                    @leastaction @xgranade it just shows that most people (represented by politicians) don’t care. They care much more about economic growth and their wealth then ethics. Also people in general are extremely lazy, which is why it is so tempting to use it and most are even willing to outsource their thinking to it.

                                    neongod@mstdn.socialN This user is from outside of this forum
                                    neongod@mstdn.socialN This user is from outside of this forum
                                    neongod@mstdn.social
                                    wrote last edited by
                                    #18

                                    @leastaction @xgranade by that I didn’t mean that politicians are not pushing their own agenda that can go against what their voters want. The only thing I claim is that the biggest flaw in any political system is ultimately people, so more democracy wouldn’t mean less push of AI.

                                    1 Reply Last reply
                                    0
                                    • xgranade@wandering.shopX xgranade@wandering.shop

                                      Yesterday, had an argument with an AI booster. I'm not going to link, both because I don't want to platform that and because I don't want anyone to go harass them. But what I thought was very interesting was that I asked point-blank if there was any degree to which ethical problems with LLMs could make them not want to use AI — they told me no, there was not, and implied that they evaluated AI purely on the basis of its efficacy.

                                      wombatpandaa@mastodon.socialW This user is from outside of this forum
                                      wombatpandaa@mastodon.socialW This user is from outside of this forum
                                      wombatpandaa@mastodon.social
                                      wrote last edited by
                                      #19

                                      @xgranade yikes...there are a lot of things that are effective at efficiency accomplishing a goal that are certainly not ethical. I understand that it's easier to rationalize away ethical concerns when it's abstracted through several layers of stubbornness, doubt, etc., but I would have a very difficult time trusting or even conversing with someone who so utterly rejects ethics as a consideration. That recent satirical post someone made about the efficient orphan smashing machine comes to mind.

                                      1 Reply Last reply
                                      0
                                      • wbftw@hachyderm.ioW wbftw@hachyderm.io

                                        @xgranade had a very similar conversation the other day too; only was able to somewhat shift my interlocutor’s position after pointing out they don’t own this “tool”, and they are at the mercy of fash/oligarch class who can (and will) start extracting rent at any moment.

                                        riotnrrd@mastodon.socialR This user is from outside of this forum
                                        riotnrrd@mastodon.socialR This user is from outside of this forum
                                        riotnrrd@mastodon.social
                                        wrote last edited by
                                        #20

                                        @wbftw @xgranade Yes! Even if we stick to one domain where “AI” works today (yes yes, FSVO), namely coding assistants, what is the future of non-commercial software if it just becomes normalised that being a programmer means paying rent of tens or hundreds of dollars a month in tokens? And that price rising once everyone is locked in, until the LLM operators can make a profit?

                                        1 Reply Last reply
                                        0
                                        • yosh@toot.yosh.isY yosh@toot.yosh.is

                                          @xgranade

                                          I'm not sure I'd say "AI doesn't work" anymore. It definitely doesn't "work" to the degree that the loudest boosters will claim it does. But like, I do think it's recently crossed a threshold where it can be a useful tool in the right hands.

                                          Which I personally find very annoying since I too have moral qualms about the broader AI industry. E.g. the point about surveillance you're making I think is an important one.

                                          mms@mastodon.bsd.cafeM This user is from outside of this forum
                                          mms@mastodon.bsd.cafeM This user is from outside of this forum
                                          mms@mastodon.bsd.cafe
                                          wrote last edited by
                                          #21

                                          @yosh @xgranade I'm on the same boat ;-(

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups