Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. the idea that a "singularity" is possible is just the idea that you can turn "mistaking a sigmoid for an exponential" into a millenarian religion

the idea that a "singularity" is possible is just the idea that you can turn "mistaking a sigmoid for an exponential" into a millenarian religion

Scheduled Pinned Locked Moved Uncategorized
67 Posts 35 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • glyph@mastodon.socialG glyph@mastodon.social

    casual thinkpieces and lazy attempts at scicomm are what has set me off but the actual thing I'm mad about is that we are ruled by people with a child's understanding of the world and the economy and that's actually really bad

    clarkiestar@mas.toC This user is from outside of this forum
    clarkiestar@mas.toC This user is from outside of this forum
    clarkiestar@mas.to
    wrote last edited by
    #42

    @glyph really good to read a sane alternative to what is usually said in the media about AI

    1 Reply Last reply
    0
    • glyph@mastodon.socialG glyph@mastodon.social

      the idea that a "singularity" is possible is just the idea that you can turn "mistaking a sigmoid for an exponential" into a millenarian religion

      ireneista@adhd.irenes.spaceI This user is from outside of this forum
      ireneista@adhd.irenes.spaceI This user is from outside of this forum
      ireneista@adhd.irenes.space
      wrote last edited by
      #43

      @glyph yeah it's the rapture for people who find computers easier to believe in than old men

      1 Reply Last reply
      0
      • glyph@mastodon.socialG glyph@mastodon.social

        RE: https://mastodon.social/@glyph/115076275195904439

        I've written about this before and I will probably do it again. but I don't know what else to do but repeat myself when allegedly serious, internationally-renowned academic experts and influential public intellectuals are just going out there and saying stuff that would get you laughed out of a late night freshman dorm room conversation about philosophy

        semanticist@mastodon.socialS This user is from outside of this forum
        semanticist@mastodon.socialS This user is from outside of this forum
        semanticist@mastodon.social
        wrote last edited by
        #44

        @glyph The only scenario I’ve found interesting is the idea that a sufficiently advanced AI doesn’t need to replace the people, just be so amazingly perceptive that it can convince, blackmail, or threaten anyone it can communicate with into doing anything it wanted.

        It’s a great idea… when I read it in 2000AD comics. But only good enough to be my third favourite series after Judge Dredd and Rogue Trooper, not something that keeps me up at night.

        1 Reply Last reply
        0
        • glyph@mastodon.socialG glyph@mastodon.social

          seriously just imagine the plot of one of the movies that doomers seem to think are documentaries, like Terminator 2. imagine the scene where the T-1000 is getting pelted with bullets. instead of seamlessly autonomously healing, imagine it has to lie down and wait for a human to place an order for $1,000,000 of NVIDIA GPUs to be delivered in a shipping container and then a construction crew to set up a methane generator to run for two weeks straight before it got up again. is that still scary?

          dabeaz@mastodon.socialD This user is from outside of this forum
          dabeaz@mastodon.socialD This user is from outside of this forum
          dabeaz@mastodon.social
          wrote last edited by
          #45

          @glyph I've seen enough movies to know that the whole thing will come crashing down due to a very tiny inconsequential unnoticed design flaw. You know, like an expired SSL certificate.

          joxn@wandering.shopJ 1 Reply Last reply
          0
          • glyph@mastodon.socialG glyph@mastodon.social

            seriously just imagine the plot of one of the movies that doomers seem to think are documentaries, like Terminator 2. imagine the scene where the T-1000 is getting pelted with bullets. instead of seamlessly autonomously healing, imagine it has to lie down and wait for a human to place an order for $1,000,000 of NVIDIA GPUs to be delivered in a shipping container and then a construction crew to set up a methane generator to run for two weeks straight before it got up again. is that still scary?

            nosword@localization.cafeN This user is from outside of this forum
            nosword@localization.cafeN This user is from outside of this forum
            nosword@localization.cafe
            wrote last edited by
            #46

            @glyph This is a great thread but it IS scary to consider that there absolutely would be police standing guard over it until it can be fixed, people saying “If we don't repair the transforming killing machine, China will,” an op-ed in the NYT headed “My Don’t-Want-To-Be-Killed-By-a-Smirking-Robert-Patrick Friends Are Crazy,” principals signing deals with Google to have murderbots stalk classrooms (guardrails: only kill kids named John Connor), &c

            1 Reply Last reply
            0
            • glyph@mastodon.socialG glyph@mastodon.social

              seriously just imagine the plot of one of the movies that doomers seem to think are documentaries, like Terminator 2. imagine the scene where the T-1000 is getting pelted with bullets. instead of seamlessly autonomously healing, imagine it has to lie down and wait for a human to place an order for $1,000,000 of NVIDIA GPUs to be delivered in a shipping container and then a construction crew to set up a methane generator to run for two weeks straight before it got up again. is that still scary?

              f4grx@chaos.socialF This user is from outside of this forum
              f4grx@chaos.socialF This user is from outside of this forum
              f4grx@chaos.social
              wrote last edited by
              #47

              @glyph skynet was so intelligent, they built terminators so efficienly, they run on bare 6502s ; they dont even need nvidia GPUs.

              LLMs are not even close.

              1 Reply Last reply
              0
              • glyph@mastodon.socialG glyph@mastodon.social

                doomers might look at my rant here and think, "but wait, once it's self-sustaining, even a little, it's TOO LATE, it's already out of control!!!" and to that I say: no. not even close. look the evolution of *any* business. managing resource flows is really hard. there is an off-ramp every single day

                glennseto@mastodon.socialG This user is from outside of this forum
                glennseto@mastodon.socialG This user is from outside of this forum
                glennseto@mastodon.social
                wrote last edited by
                #48

                @glyph Another counterpoint: Every single zombie apocalypse scenario, where the collapse of human infrastructure and supply chains is so absolute, not even the zombies disappearing overnight would still lead to years, if not decades of recovery.

                1 Reply Last reply
                0
                • glyph@mastodon.socialG glyph@mastodon.social

                  if, in order to achieve your out-of-control doomsday robot scenario, a trillion dollars worth of human effort must be expended annually, and if any of it stops for even a moment than the whole thing implodes and grinds to a halt, _you can stop worrying_ that it is "the machines" which dominate us

                  ced@mapstodon.spaceC This user is from outside of this forum
                  ced@mapstodon.spaceC This user is from outside of this forum
                  ced@mapstodon.space
                  wrote last edited by
                  #49

                  @glyph above all, if people believe singularity is scary, why the fuck do they invest a trillion $/yr to try to reach it ? At that point, we should try to convince them that a bigger CERN could really provoke a black hole on earth.
                  Won’t work either, but at least we’ll have something useful at the end for a fraction of the price!

                  1 Reply Last reply
                  0
                  • glyph@mastodon.socialG glyph@mastodon.social

                    like if anyone had halfway-plausible "grey goo" nanotech that could do anything that looked like computation, that might be worrying. a locally viable self-reproducing platform that can make another one of itself from a pile of dirt, even if it's like, special dirt, that might scare me a little bit. but an overlord hive-mind that requires an uninterrupted global high-purity helium supply chain just to make ONE more of itself is supposed to be a threat?

                    glennseto@mastodon.socialG This user is from outside of this forum
                    glennseto@mastodon.socialG This user is from outside of this forum
                    glennseto@mastodon.social
                    wrote last edited by
                    #50

                    @glyph Goddammit, this is twice in a row I'm forced to root for, of all things, the government of Iran.

                    Edit: For context, a lot of the world's helium trade goes through, you guessed it, the Strait of Hormuz.

                    1 Reply Last reply
                    0
                    • glyph@mastodon.socialG glyph@mastodon.social

                      the idea that a "singularity" is possible is just the idea that you can turn "mistaking a sigmoid for an exponential" into a millenarian religion

                      ahltorp@mastodon.nuA This user is from outside of this forum
                      ahltorp@mastodon.nuA This user is from outside of this forum
                      ahltorp@mastodon.nu
                      wrote last edited by
                      #51

                      @glyph Believing LLM chatbots will achieve singularity is like someone believing teleportation and manufacture-anything-machines are right around the corner because they once saw a magician perform a magic trick.

                      1 Reply Last reply
                      0
                      • mcc@mastodon.socialM mcc@mastodon.social

                        @glyph my assertion was that the singularity, as described by ray kurzweil, accurately describes the invention of writing, and i don't see why it would be more interesting if the self-improving intelligent mechanism were made of etched silicon instead of CHNOPS nanomachines. it is harder for etched silicon to self-reproduce, anyway. the CHNOPS nanomachines just do that.

                        i think human advancement *has* followed an exponential-*looking* curve since that point, albeit with a low base.

                        lockex@ioc.exchangeL This user is from outside of this forum
                        lockex@ioc.exchangeL This user is from outside of this forum
                        lockex@ioc.exchange
                        wrote last edited by
                        #52

                        @mcc @glyph
                        Language is a virus https://en.wikipedia.org/wiki/Language_is_a_Virus

                        1 Reply Last reply
                        0
                        • glyph@mastodon.socialG glyph@mastodon.social

                          in order to be a singularity candidate, an AI would need to achieve vertical integration from silicon fabrication through logistics and integration, into operating systems and applications, with tight whole-system feedback from the robotics to the shipping to the power generation and back

                          varx@cybersecurity.theaterV This user is from outside of this forum
                          varx@cybersecurity.theaterV This user is from outside of this forum
                          varx@cybersecurity.theater
                          wrote last edited by
                          #53

                          @glyph This *strongly* depends on what you mean by "singularity". I think you're conflating that with "hard takeoff paperclips scenario" or something.

                          I can just barely (barely!) imagine a future where someone manages to use AI to get a more efficient form of AI, which would allow further bootstrapping without requiring more hardware. Same hardware gets more compute.

                          You're spot-on about the supply chain limitations, though. Good luck to the AI that wants to dig up more cobalt or whatever.

                          glyph@mastodon.socialG 1 Reply Last reply
                          0
                          • glyph@mastodon.socialG glyph@mastodon.social

                            casual thinkpieces and lazy attempts at scicomm are what has set me off but the actual thing I'm mad about is that we are ruled by people with a child's understanding of the world and the economy and that's actually really bad

                            mathaetaes@infosec.exchangeM This user is from outside of this forum
                            mathaetaes@infosec.exchangeM This user is from outside of this forum
                            mathaetaes@infosec.exchange
                            wrote last edited by
                            #54

                            @glyph I’m not an AI doomer and I pretty much agree with your whole thread, but something to consider: when I think of AI taking over humanity, I see The Matrix more than Terminator. A hypothetical super intelligent AI wouldn’t necessarily need complete self-replication ability… it just needs the ability to influence humans enough to have them do the parts it can’t.

                            If you think about it, we’ve been building tech designed specifically to manipulate human behavior since the advent of social media… and it’s effective. If humanity were any good at protecting itself or organizing for the greater good, the perverse reward systems of capitalism would’ve been brought into check long before US oligarchs like Musk and Zuck could have been able to amass their power.

                            In a hypothetical world where machines rule, the more likely scenario is a majority of humans self-oppressing because they’ve been manipulated into it by adjustments to the algorithms that feed them the information they use to establish reality. We become part of the system.

                            The only real difference between that world and today is that today it’s a handful of billionaires controlling the algorithms. Replace Zuck and a few others with sufficiently capable AI and is it really that unbelievable that society would just keep cranking out more machines despite a slow degradation of quality of life?

                            Anyway - great thread and a fun topic to kick around. Thanks for posting it.

                            dalias@hachyderm.ioD 1 Reply Last reply
                            0
                            • glyph@mastodon.socialG glyph@mastodon.social

                              seriously just imagine the plot of one of the movies that doomers seem to think are documentaries, like Terminator 2. imagine the scene where the T-1000 is getting pelted with bullets. instead of seamlessly autonomously healing, imagine it has to lie down and wait for a human to place an order for $1,000,000 of NVIDIA GPUs to be delivered in a shipping container and then a construction crew to set up a methane generator to run for two weeks straight before it got up again. is that still scary?

                              fxchip@hachyderm.ioF This user is from outside of this forum
                              fxchip@hachyderm.ioF This user is from outside of this forum
                              fxchip@hachyderm.io
                              wrote last edited by
                              #55

                              @glyph This is all a component I hadn't really considered but entirely makes sense. And simultaneously reminds me of that scene in Ocean's Eleven where the one guy is trying to tell everybody why robbing a casino is a stupid fucking idea, capping the whole thing off with "and even if you *do* manage to pull all that off, *you're still in the middle of the fucking desert!*" (Which, tangentially, has become something of a metaphorical benchmark for me deciding whether it's worth doing some huge labor in general)

                              All that to say, awesome thread 😄

                              1 Reply Last reply
                              0
                              • glyph@mastodon.socialG glyph@mastodon.social

                                @suetanvil it's ruining my ability to appreciate the fantasy!!!

                                suetanvil@freeradical.zoneS This user is from outside of this forum
                                suetanvil@freeradical.zoneS This user is from outside of this forum
                                suetanvil@freeradical.zone
                                wrote last edited by
                                #56

                                @glyph

                                That 'supposedly' is the clue.

                                (My guess: they are hardcore rationalists who took that position in order to flatter their own egos ("*I'm* too *smart* to believe in the Sky Bully") and so didn't think that position through. They are also so insulated from reality by their wealth that they don't know how to handle being wrong about something. So now they deal with the downsides of that choice by techbro-inventing religion from scrstch^W SF.)

                                suetanvil@freeradical.zoneS 1 Reply Last reply
                                0
                                • suetanvil@freeradical.zoneS suetanvil@freeradical.zone

                                  @glyph

                                  That 'supposedly' is the clue.

                                  (My guess: they are hardcore rationalists who took that position in order to flatter their own egos ("*I'm* too *smart* to believe in the Sky Bully") and so didn't think that position through. They are also so insulated from reality by their wealth that they don't know how to handle being wrong about something. So now they deal with the downsides of that choice by techbro-inventing religion from scrstch^W SF.)

                                  suetanvil@freeradical.zoneS This user is from outside of this forum
                                  suetanvil@freeradical.zoneS This user is from outside of this forum
                                  suetanvil@freeradical.zone
                                  wrote last edited by
                                  #57

                                  @glyph

                                  (A techbro would rather sacrifice the world on the altar of his dark GPU gods than go to therapy.)

                                  1 Reply Last reply
                                  0
                                  • dabeaz@mastodon.socialD dabeaz@mastodon.social

                                    @glyph I've seen enough movies to know that the whole thing will come crashing down due to a very tiny inconsequential unnoticed design flaw. You know, like an expired SSL certificate.

                                    joxn@wandering.shopJ This user is from outside of this forum
                                    joxn@wandering.shopJ This user is from outside of this forum
                                    joxn@wandering.shop
                                    wrote last edited by
                                    #58

                                    @dabeaz @glyph leap second goes the wrong way

                                    joxn@wandering.shopJ 1 Reply Last reply
                                    0
                                    • joxn@wandering.shopJ joxn@wandering.shop

                                      @dabeaz @glyph leap second goes the wrong way

                                      joxn@wandering.shopJ This user is from outside of this forum
                                      joxn@wandering.shopJ This user is from outside of this forum
                                      joxn@wandering.shop
                                      wrote last edited by
                                      #59

                                      @dabeaz @glyph Bruce Willis, Sylvester Stallone, and Ahnold team up to travel to the earth’s core and explode a nuclear bomb, causing the earth’s rotation to speed up enough to result in an extra negative leap second

                                      dabeaz@mastodon.socialD 1 Reply Last reply
                                      0
                                      • joxn@wandering.shopJ joxn@wandering.shop

                                        @dabeaz @glyph Bruce Willis, Sylvester Stallone, and Ahnold team up to travel to the earth’s core and explode a nuclear bomb, causing the earth’s rotation to speed up enough to result in an extra negative leap second

                                        dabeaz@mastodon.socialD This user is from outside of this forum
                                        dabeaz@mastodon.socialD This user is from outside of this forum
                                        dabeaz@mastodon.social
                                        wrote last edited by
                                        #60

                                        @joXn @glyph But all for naught due to some bullshit involving an unpaid invoice at Network Solutions.

                                        1 Reply Last reply
                                        0
                                        • varx@cybersecurity.theaterV varx@cybersecurity.theater

                                          @glyph This *strongly* depends on what you mean by "singularity". I think you're conflating that with "hard takeoff paperclips scenario" or something.

                                          I can just barely (barely!) imagine a future where someone manages to use AI to get a more efficient form of AI, which would allow further bootstrapping without requiring more hardware. Same hardware gets more compute.

                                          You're spot-on about the supply chain limitations, though. Good luck to the AI that wants to dig up more cobalt or whatever.

                                          glyph@mastodon.socialG This user is from outside of this forum
                                          glyph@mastodon.socialG This user is from outside of this forum
                                          glyph@mastodon.social
                                          wrote last edited by
                                          #61

                                          @varx imagining stuff is easy though. I can imagine lots of stuff

                                          1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups