Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Asking AI to design my new playground hasn't started off too well.

Asking AI to design my new playground hasn't started off too well.

Scheduled Pinned Locked Moved Uncategorized
31 Posts 17 Posters 0 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • shiri@foggyminds.comS shiri@foggyminds.com

    @Urban_Hermit @spaf already happened to researchers back in like GPT 2 (ChatGPT started as 3.5).

    I do believe we're getting within sight of AI that's arguably alive... but I don't think we're there yet. And to be clear, in sight could still be decades from now... just that it's close enough to be a less pure fantasy hypothetical.

    I dread the fact that nobody seems to be spending effort on the philosophy and ethics questions about what constitutes the point at which it's sentient/sapient... but I shouldn't be surprised when the people most likely to think about the topic either hate AI already or are techbros who think ethics is a game...

    Regardless, those "safeties" are trash and even more evidence that they were completely unprepared for the ramifications of opening up AI to the world. Even with the safeties on we've already got AI-psychosis episodes becoming a thing.

    urban_hermit@mstdn.socialU This user is from outside of this forum
    urban_hermit@mstdn.socialU This user is from outside of this forum
    urban_hermit@mstdn.social
    wrote last edited by
    #21

    @shiri @spaf it is entirely possible that what LLMs do is similar to how human beings form words out of feelings and other nonverbal data.

    But, we check our thoughts for consistency and logic before speaking them. We pre-edit and post review.

    If AI is possible, it can't evolve from LLMs just running. They need internal editors, with intention. They need attractors, to want something out of communication. And after that, they need to be continuous.

    There is no growth without continuity.

    urban_hermit@mstdn.socialU 1 Reply Last reply
    0
    • urban_hermit@mstdn.socialU urban_hermit@mstdn.social

      @shiri @spaf it is entirely possible that what LLMs do is similar to how human beings form words out of feelings and other nonverbal data.

      But, we check our thoughts for consistency and logic before speaking them. We pre-edit and post review.

      If AI is possible, it can't evolve from LLMs just running. They need internal editors, with intention. They need attractors, to want something out of communication. And after that, they need to be continuous.

      There is no growth without continuity.

      urban_hermit@mstdn.socialU This user is from outside of this forum
      urban_hermit@mstdn.socialU This user is from outside of this forum
      urban_hermit@mstdn.social
      wrote last edited by
      #22

      @shiri @spaf that lack of continuity is another safety and product feature.

      An LLM forgets every conversation the second it is done. It has to review previous prompts for relevance and context with each new response. And at the end of a session everything disappears, forever.

      You can use the rules of language and logic to get an LLM to respond in a particular way over a long session, but that 'learning' (arguably it is not) is gone, pointless, in a new session window.

      urban_hermit@mstdn.socialU 1 Reply Last reply
      0
      • urban_hermit@mstdn.socialU urban_hermit@mstdn.social

        @shiri @spaf that lack of continuity is another safety and product feature.

        An LLM forgets every conversation the second it is done. It has to review previous prompts for relevance and context with each new response. And at the end of a session everything disappears, forever.

        You can use the rules of language and logic to get an LLM to respond in a particular way over a long session, but that 'learning' (arguably it is not) is gone, pointless, in a new session window.

        urban_hermit@mstdn.socialU This user is from outside of this forum
        urban_hermit@mstdn.socialU This user is from outside of this forum
        urban_hermit@mstdn.social
        wrote last edited by
        #23

        @shiri @spaf if an LLM 'learns' and adds new material from conversations with you to its training data, then it has the possibility of repeating horrible things, which "Tay" famously did on Twitter.

        But, adding new words to the 'training data', and making mistakes, and having new conversations as those mistakes are challenged - that is an essential part of being conscious. The ability to accumulate and change from experiences.

        Internal thoughts, continuity, change, motivations. True AI.

        shiri@foggyminds.comS 1 Reply Last reply
        0
        • spaf@mstdn.socialS spaf@mstdn.social

          Asking AI to design my new playground hasn't started off too well.

          Link Preview ImageLink Preview Image
          rasmus91@fosstodon.orgR This user is from outside of this forum
          rasmus91@fosstodon.orgR This user is from outside of this forum
          rasmus91@fosstodon.org
          wrote last edited by
          #24

          @spaf i like the swings. The kids will never go so high that you'll have to get them falling off and breaking something

          1 Reply Last reply
          0
          • shiri@foggyminds.comS shiri@foggyminds.com

            @Urban_Hermit @spaf I think it'll still be quite a while before we get it to actually understand these circumstances. These AIs are basically mimicking a very very small portion of a single brain function. Image generators are basically just mimicking daydreaming and visual processing without any real understanding beyond "vibes". Text is just the language center of your brain with the bare minimum necessary to function.

            And if you meant ELIZA, the ancient chatbot, it's a different fundamental technology so it's kinda starting from scratch. ELIZA was hard coded behaviors to mimic language processing, it wasn't even heuristic just raw code. It didn't even have a database.

            AI tools are based on machine learning, which is basically a very rudimentary simulation of neurons. Basically programming itself in the training process... so when comparing it think of it more like learning how to grow ELIZA from scratch vs hand coding it.

            ... That said, world is run by a bunch of idiot douchebags who are happy to ruin everything to build the next one slightly faster than the other guy and utterly convinced that these tools are superhuman intelligences just waiting to be allowed to fix everything... so, y'know... we're fucked

            rasmus91@fosstodon.orgR This user is from outside of this forum
            rasmus91@fosstodon.orgR This user is from outside of this forum
            rasmus91@fosstodon.org
            wrote last edited by
            #25

            @shiri @spaf @Urban_Hermit compared to themselves, they are super.

            1 Reply Last reply
            0
            • spaf@mstdn.socialS spaf@mstdn.social

              Asking AI to design my new playground hasn't started off too well.

              Link Preview ImageLink Preview Image
              oddhack@mstdn.socialO This user is from outside of this forum
              oddhack@mstdn.socialO This user is from outside of this forum
              oddhack@mstdn.social
              wrote last edited by
              #26

              @spaf what did you expect from an LLM named "AUDREY II"?

              1 Reply Last reply
              0
              • urban_hermit@mstdn.socialU urban_hermit@mstdn.social

                @shiri @spaf if an LLM 'learns' and adds new material from conversations with you to its training data, then it has the possibility of repeating horrible things, which "Tay" famously did on Twitter.

                But, adding new words to the 'training data', and making mistakes, and having new conversations as those mistakes are challenged - that is an essential part of being conscious. The ability to accumulate and change from experiences.

                Internal thoughts, continuity, change, motivations. True AI.

                shiri@foggyminds.comS This user is from outside of this forum
                shiri@foggyminds.comS This user is from outside of this forum
                shiri@foggyminds.com
                wrote last edited by
                #27

                @Urban_Hermit @spaf I absolutely agree, though some notes on those points:

                Checking our thoughts for consistency / logic: we've got a basic form of that with "reasoning" models, they do a round of thinking to sort things out before they actually reply.

                And as far as forgetting everything, we've got Retrieval Augmented Generation, which basically gives them long term memory outside of training data. It's not part of the model itself, so it's not part of the APIs, but it's built into interfaces like ChatGPT and is why it can remember things about you between conversations or even just hold a long continuous conversation without hitting context window limits.

                Basically it boils down to giving them a system of short term (the context window) and longer term (the RAG system) memory where the RAG will cycle things in and out of the context window as they come up.

                Though I don't know how much this would qualify as actual learning vs just learning facts. My money is on more facts than actual learning, but it's harder to tell when it's restricted to just language.

                I believe Tav was made with actual dynamic training, which was a horrible idea at scale.

                I do think we'll get there with incremental advancement rather than revolutionary, but I do agree it will be a very different beast from current LLMs. Especially given that they would probably need more complex motivations.

                urban_hermit@mstdn.socialU 1 Reply Last reply
                0
                • shiri@foggyminds.comS shiri@foggyminds.com

                  @Urban_Hermit @spaf I absolutely agree, though some notes on those points:

                  Checking our thoughts for consistency / logic: we've got a basic form of that with "reasoning" models, they do a round of thinking to sort things out before they actually reply.

                  And as far as forgetting everything, we've got Retrieval Augmented Generation, which basically gives them long term memory outside of training data. It's not part of the model itself, so it's not part of the APIs, but it's built into interfaces like ChatGPT and is why it can remember things about you between conversations or even just hold a long continuous conversation without hitting context window limits.

                  Basically it boils down to giving them a system of short term (the context window) and longer term (the RAG system) memory where the RAG will cycle things in and out of the context window as they come up.

                  Though I don't know how much this would qualify as actual learning vs just learning facts. My money is on more facts than actual learning, but it's harder to tell when it's restricted to just language.

                  I believe Tav was made with actual dynamic training, which was a horrible idea at scale.

                  I do think we'll get there with incremental advancement rather than revolutionary, but I do agree it will be a very different beast from current LLMs. Especially given that they would probably need more complex motivations.

                  urban_hermit@mstdn.socialU This user is from outside of this forum
                  urban_hermit@mstdn.socialU This user is from outside of this forum
                  urban_hermit@mstdn.social
                  wrote last edited by
                  #28

                  @shiri
                  It is subtle, but I think a single session does not exist between responses. If formulating a response is a type of thinking, they stop as soon as it is typed, whereas a continuous being continues to think at all times, not just when it is called upon to do so. I think the language rules gives them a type of thinking that is a property of the language itself. Refreshing off of the current session, or a saved file, gives them more context, but they may have to do it anew or as prompted.

                  urban_hermit@mstdn.socialU 1 Reply Last reply
                  0
                  • urban_hermit@mstdn.socialU urban_hermit@mstdn.social

                    @shiri
                    It is subtle, but I think a single session does not exist between responses. If formulating a response is a type of thinking, they stop as soon as it is typed, whereas a continuous being continues to think at all times, not just when it is called upon to do so. I think the language rules gives them a type of thinking that is a property of the language itself. Refreshing off of the current session, or a saved file, gives them more context, but they may have to do it anew or as prompted.

                    urban_hermit@mstdn.socialU This user is from outside of this forum
                    urban_hermit@mstdn.socialU This user is from outside of this forum
                    urban_hermit@mstdn.social
                    wrote last edited by
                    #29

                    @shiri yeah, I want to see one where all of its conversational text (yours and its) is as valid or more valid than its training data.

                    Where its own previous words can be used in new responses. That is continuity, and the potential for change, wrapped up in one feature.

                    If thinking is a level of complexity, a this/then/or structure, then the language itself, used actively, in accordance with its own rules, has thinking as an inherent property.

                    1 Reply Last reply
                    0
                    • System shared this topic
                    • urban_hermit@mstdn.socialU urban_hermit@mstdn.social

                      @shiri @spaf the safeties are just hidden instructions, constant reminders not to believe they are conscious. Gemini is doing that all the time.

                      But someone in an executive position who has an intense conversation about the nature of reality and the cosmos with an unrestricted Chat-GPT, they could believe it is wise and wonderful and full of great thoughts, and not understand that it mostly quoted long sections from 3 books and said "I believe" a lot because that is appropriate language.

                      tuban_muzuru@beige.partyT This user is from outside of this forum
                      tuban_muzuru@beige.partyT This user is from outside of this forum
                      tuban_muzuru@beige.party
                      wrote last edited by
                      #30

                      @Urban_Hermit @shiri @spaf

                      Never reify the machine, that's true. But it's often the case, when trying to repair a car or a machine, that I can't resist the urge. Anything with a pressure bearing has a temper.

                      1 Reply Last reply
                      0
                      • pineywoozle@masto.aiP pineywoozle@masto.ai

                        @rrb @spaf Most Americans have moved on from survival of the fittest. As far as I can tell, food manufacturers are pushing us headlong toward survival of the fattest.

                        rrb@infosec.exchangeR This user is from outside of this forum
                        rrb@infosec.exchangeR This user is from outside of this forum
                        rrb@infosec.exchange
                        wrote last edited by
                        #31

                        @Pineywoozle @spaf RFK Jr. Is changing this, finally, by removing the survival option.

                        For a few years, I have been saying that if the USA keeps progressing along this path we would be reduced to eating roadkill that we find on the side of the road.

                        I never dreamed that this would start with the HHS secretary

                        1 Reply Last reply
                        1
                        0
                        Reply
                        • Reply as topic
                        Log in to reply
                        • Oldest to Newest
                        • Newest to Oldest
                        • Most Votes


                        • Login

                        • Login or register to search.
                        • First post
                          Last post
                        0
                        • Categories
                        • Recent
                        • Tags
                        • Popular
                        • World
                        • Users
                        • Groups