Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Writing this up again so I can pin it: AI is literally a fascist project.

Writing this up again so I can pin it: AI is literally a fascist project.

Scheduled Pinned Locked Moved Uncategorized
40 Posts 7 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • jens@social.finkhaeuser.deJ jens@social.finkhaeuser.de

    @GyrosGeier I actually find it difficult to write good image descriptions. The ones I write zero in on the point I want to make, but often omit details. In a way, that's a writing faux pas. In creative writing you learn "show, don't tell", and I do the opposite.

    This isn't a counter-argument (nor an argument). All I want to do is acknowledge how hard it is to do well with assitance of this kind.

    gyrosgeier@hachyderm.ioG This user is from outside of this forum
    gyrosgeier@hachyderm.ioG This user is from outside of this forum
    gyrosgeier@hachyderm.io
    wrote last edited by
    #31

    @jens that is a good description though: the details aren't important, but the point is. If you can't show because the recipient is vision impaired, then you need to tell.

    My point is that while AI has its uses in assistive technologies, it is also inherently limited, so it's not a good direction to take research in assistive technologies in.

    1 Reply Last reply
    0
    • jens@social.finkhaeuser.deJ jens@social.finkhaeuser.de

      ... scarcity, in which - by whichever proof scheme - those who participate early in the system benefit off those who come later (aka pyramid schemes). The proof algorithm guarantees scarcity; it's the whole point of blockchain vs. any other distributed system that there is a chokehold on resource creation somewhere.

      AI is doing much the same thing, but it doesn't advertise this artificial scarcity as part of the solution. Instead, it simply guarantees that those who already own the most...

      nielsa@mas.toN This user is from outside of this forum
      nielsa@mas.toN This user is from outside of this forum
      nielsa@mas.to
      wrote last edited by
      #32

      @jens The way the global stock market works is an interesting progenitor for cryptocurrencies, too. It used to be traded mostly based on earnings paid for holding the stock, but has in recent decades transitioned into being traded speculatively, which makes each stock into its own little proto-ponzi scheme.

      jens@social.finkhaeuser.deJ 1 Reply Last reply
      0
      • jens@social.finkhaeuser.deJ jens@social.finkhaeuser.de

        @nielsa I think you need to read the entire thread 🙂

        nielsa@mas.toN This user is from outside of this forum
        nielsa@mas.toN This user is from outside of this forum
        nielsa@mas.to
        wrote last edited by
        #33

        @jens I read the thread, it's a good thread.

        I guess I'm just delineating the caveat of what kind of LLM can be neutral technology. Which *is* a minor footnote in what is currently happening.

        Thanks for writing this up 😁

        jens@social.finkhaeuser.deJ 1 Reply Last reply
        0
        • nielsa@mas.toN nielsa@mas.to

          @jens The way the global stock market works is an interesting progenitor for cryptocurrencies, too. It used to be traded mostly based on earnings paid for holding the stock, but has in recent decades transitioned into being traded speculatively, which makes each stock into its own little proto-ponzi scheme.

          jens@social.finkhaeuser.deJ This user is from outside of this forum
          jens@social.finkhaeuser.deJ This user is from outside of this forum
          jens@social.finkhaeuser.de
          wrote last edited by
          #34

          @nielsa Oh, yes.

          My understanding of financial products isn't exactly complete, but my take is that they all fall into two categories.

          I mean, buying stock is a bet on future earnings. You can lose that bet, so one category is to aggregate things in such a way that - hopefully - losses in one are offset by gains in the other.

          The other category is a layer of indirection, i.e. bets on something other people are betting on.

          All of this multi-layered to the point where you can't know what...

          jens@social.finkhaeuser.deJ 1 Reply Last reply
          0
          • jens@social.finkhaeuser.deJ jens@social.finkhaeuser.de

            @nielsa Oh, yes.

            My understanding of financial products isn't exactly complete, but my take is that they all fall into two categories.

            I mean, buying stock is a bet on future earnings. You can lose that bet, so one category is to aggregate things in such a way that - hopefully - losses in one are offset by gains in the other.

            The other category is a layer of indirection, i.e. bets on something other people are betting on.

            All of this multi-layered to the point where you can't know what...

            jens@social.finkhaeuser.deJ This user is from outside of this forum
            jens@social.finkhaeuser.deJ This user is from outside of this forum
            jens@social.finkhaeuser.de
            wrote last edited by
            #35

            @nielsa ... you're betting on, which makes ponzi schemes and insider trading so much more effective, as the costs are externalized to the average shareholder.

            And people think this is serious business.

            The only thing that seems serious about it is that it seriously affects us.

            1 Reply Last reply
            0
            • nielsa@mas.toN nielsa@mas.to

              @jens I read the thread, it's a good thread.

              I guess I'm just delineating the caveat of what kind of LLM can be neutral technology. Which *is* a minor footnote in what is currently happening.

              Thanks for writing this up 😁

              jens@social.finkhaeuser.deJ This user is from outside of this forum
              jens@social.finkhaeuser.deJ This user is from outside of this forum
              jens@social.finkhaeuser.de
              wrote last edited by
              #36

              @nielsa And frankly, as a neutral tech or tool, I do find the whole thing interesting!

              It's just... pretty much like fusion is interesting. I would love for us to have cheap, safe "desktop" fusion.

              It's just always been 20 years away, and inextricably tied up with dirty fission, so how can one *practically* support one and not the other?

              The cost-benefit-analysis suggests to me that the cost of getting this wrong is so much higher than the cost of missing out on good stuff, though.

              nielsa@mas.toN 1 Reply Last reply
              0
              • jens@social.finkhaeuser.deJ jens@social.finkhaeuser.de

                @nielsa And frankly, as a neutral tech or tool, I do find the whole thing interesting!

                It's just... pretty much like fusion is interesting. I would love for us to have cheap, safe "desktop" fusion.

                It's just always been 20 years away, and inextricably tied up with dirty fission, so how can one *practically* support one and not the other?

                The cost-benefit-analysis suggests to me that the cost of getting this wrong is so much higher than the cost of missing out on good stuff, though.

                nielsa@mas.toN This user is from outside of this forum
                nielsa@mas.toN This user is from outside of this forum
                nielsa@mas.to
                wrote last edited by
                #37

                @jens Absolutely agree on all of that.

                I have a few ideas I think could make good, ethical use of generalized LLMs, but only assuming no side benefits to the people largely driving their development and to some extent that the LLM itself is produced ethically... and that leaves a very narrow space and thus a significant startup cost...

                1 Reply Last reply
                0
                • jens@social.finkhaeuser.deJ jens@social.finkhaeuser.de

                  Writing this up again so I can pin it: AI is literally a fascist project. Friends don't let friends use it.

                  Before I go into this, there are two types of responses to this that I have taken seriously so far.

                  One I'll call HashTagNotAllAI, which yields the obligatory "sure", but has the same smell. I'll leave it at that.

                  The other is that an anti AI stance also throws some assistive technology under the bus, making such a stance intrinsically ableistic. The easy thing to do is to refer...

                  lwriemen@social.librem.oneL This user is from outside of this forum
                  lwriemen@social.librem.oneL This user is from outside of this forum
                  lwriemen@social.librem.one
                  wrote last edited by
                  #38

                  @jens AI is too confusing of a term, especially when talking about assistance. e.g., can text to speech or voice recognition technology be called AI? It certainly doesn't a rainforest destroying LLM level of technology; it's been around for at least 35 years.

                  I don't stay abreast of all the assistive technology, but is there any that really requires LLMs at massive scale?

                  jens@social.finkhaeuser.deJ mcv@friendica.opensocial.spaceM 2 Replies Last reply
                  0
                  • lwriemen@social.librem.oneL lwriemen@social.librem.one

                    @jens AI is too confusing of a term, especially when talking about assistance. e.g., can text to speech or voice recognition technology be called AI? It certainly doesn't a rainforest destroying LLM level of technology; it's been around for at least 35 years.

                    I don't stay abreast of all the assistive technology, but is there any that really requires LLMs at massive scale?

                    jens@social.finkhaeuser.deJ This user is from outside of this forum
                    jens@social.finkhaeuser.deJ This user is from outside of this forum
                    jens@social.finkhaeuser.de
                    wrote last edited by
                    #39

                    @lwriemen As has been mentioned in a sub-thread, there e.g. exist things that analyze an image and provide textual desceiptions.

                    In the broader sense, translation is an assistive tech for non-native speakers of any language.

                    1 Reply Last reply
                    0
                    • lwriemen@social.librem.oneL lwriemen@social.librem.one

                      @jens AI is too confusing of a term, especially when talking about assistance. e.g., can text to speech or voice recognition technology be called AI? It certainly doesn't a rainforest destroying LLM level of technology; it's been around for at least 35 years.

                      I don't stay abreast of all the assistive technology, but is there any that really requires LLMs at massive scale?

                      mcv@friendica.opensocial.spaceM This user is from outside of this forum
                      mcv@friendica.opensocial.spaceM This user is from outside of this forum
                      mcv@friendica.opensocial.space
                      wrote last edited by
                      #40

                      @lwriemen @jens

                      Yes. AI is a far older and broader field than just the current LLM hype. Speech recognition, handwriting recognition, chess playing, various types of expert systems, route-finding, etc.

                      But LLMs and other modern genAI does feel different to a lot of people. And it uses a lot more data and resources.

                      1 Reply Last reply
                      0
                      • R relay@relay.mycrowd.ca shared this topic
                      Reply
                      • Reply as topic
                      Log in to reply
                      • Oldest to Newest
                      • Newest to Oldest
                      • Most Votes


                      • Login

                      • Login or register to search.
                      • First post
                        Last post
                      0
                      • Categories
                      • Recent
                      • Tags
                      • Popular
                      • World
                      • Users
                      • Groups