Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. A woman sues her insurance company for terminating her disability benefits.

A woman sues her insurance company for terminating her disability benefits.

Scheduled Pinned Locked Moved Uncategorized
66 Posts 23 Posters 2 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • falcennial@mastodon.socialF falcennial@mastodon.social

    @GyrosGeier @mjd torturous interference

    gyrosgeier@hachyderm.ioG This user is from outside of this forum
    gyrosgeier@hachyderm.ioG This user is from outside of this forum
    gyrosgeier@hachyderm.io
    wrote last edited by
    #31

    @falcennial @mjd I mean, because running an AI model is called "inference."

    1 Reply Last reply
    0
    • falcennial@mastodon.socialF falcennial@mastodon.social

      @GyrosGeier @mjd torturous interference

      mjd@mathstodon.xyzM This user is from outside of this forum
      mjd@mathstodon.xyzM This user is from outside of this forum
      mjd@mathstodon.xyz
      wrote last edited by
      #32

      @falcennial @GyrosGeier They're all closely related. They're from the Latin verb “to twist”.

      1 Reply Last reply
      0
      • divverent@social.vivaldi.netD divverent@social.vivaldi.net

        @mjd TBH I do not think OpenAI should be responsible. They're just providing a fancy random text generator to the public. And it's outright impossible to teach a random text generator to _not_ output a specific kind of text, as whatever you do, there is a way around it.

        The woman should pay all costs, as per the usual "vexatious filings" or "frivolous lawsuits" standards.

        Plus, the law in her state against practicing law without a license starts with "No person shall...". ChatGPT isn't a person.

        jonoleth@mastodon.socialJ This user is from outside of this forum
        jonoleth@mastodon.socialJ This user is from outside of this forum
        jonoleth@mastodon.social
        wrote last edited by
        #33

        @divVerent @mjd ChatGPT is not a person, which is why ChatGPT is not being sued. OpenAI sells a tool that gave her legal advice, and they certainly didn't say anywhere that it's actually just a "fancy random text generator"

        divverent@social.vivaldi.netD wellsitegeo@masto.aiW 2 Replies Last reply
        0
        • divverent@social.vivaldi.netD divverent@social.vivaldi.net

          @mjd TBH I do not think OpenAI should be responsible. They're just providing a fancy random text generator to the public. And it's outright impossible to teach a random text generator to _not_ output a specific kind of text, as whatever you do, there is a way around it.

          The woman should pay all costs, as per the usual "vexatious filings" or "frivolous lawsuits" standards.

          Plus, the law in her state against practicing law without a license starts with "No person shall...". ChatGPT isn't a person.

          mjd@mathstodon.xyzM This user is from outside of this forum
          mjd@mathstodon.xyzM This user is from outside of this forum
          mjd@mathstodon.xyz
          wrote last edited by
          #34

          @divVerent Except that there are laws against providing bogus legal advice to people, to prevent exactly this sort of situation.

          And, as you pointed out, it was OpenAI, not ChatGPT, providing the advice.

          1 Reply Last reply
          0
          • mjd@mathstodon.xyzM mjd@mathstodon.xyz

            @Infoseepage You made that up out of your head to come to the conclusion you selected beforehand.

            I don't know what actually happened, and neither do you.

            infoseepage@mastodon.socialI This user is from outside of this forum
            infoseepage@mastodon.socialI This user is from outside of this forum
            infoseepage@mastodon.social
            wrote last edited by
            #35

            @mjd Nope, I don't know the particulars of the case, but I do have some experience with the ruthlessness of insurance companies when it comes to disabled people, including a friend who went blind from cancer as a child and another who was born with severe cerebral palsy.

            1 Reply Last reply
            0
            • falcennial@mastodon.socialF falcennial@mastodon.social

              @mjd @krupo it will be a financial bubble pop, followed by what we will call 'the AI recession,' then limited, appropriate use. the shit is the dot com and GFC playbook all day.

              falcennial@mastodon.socialF This user is from outside of this forum
              falcennial@mastodon.socialF This user is from outside of this forum
              falcennial@mastodon.social
              wrote last edited by
              #36

              @mjd @krupo that MIT article demonstrated 5% of AI implementations are profit making. 95% are loss making. so when the investment goldrush mania ends or winds down, 95% of invested amounts will be wiped out. it's billions so it will have a disruptive and negative economic effect that I think we will likely experience as recession. and then that 5% of profit making implementations will be what carries forward, with further investment being modelled on those (anyone can do that right now).

              wellsitegeo@masto.aiW 1 Reply Last reply
              0
              • divverent@social.vivaldi.netD divverent@social.vivaldi.net

                @sabik @mjd It did probably exactly what she asked for in the prompt, so where's the problem? Definitely "useful".

                In case she was misinformed by ChatGPT and has to pay penalties for that reason, then _she_ should be the one suing OpenAI, not the insurance company.

                sabik@rants.auS This user is from outside of this forum
                sabik@rants.auS This user is from outside of this forum
                sabik@rants.au
                wrote last edited by
                #37

                @divVerent @mjd
                If ChatGPT misinformed her, that's not very useful

                divverent@social.vivaldi.netD 1 Reply Last reply
                0
                • jonoleth@mastodon.socialJ jonoleth@mastodon.social

                  @divVerent @mjd ChatGPT is not a person, which is why ChatGPT is not being sued. OpenAI sells a tool that gave her legal advice, and they certainly didn't say anywhere that it's actually just a "fancy random text generator"

                  divverent@social.vivaldi.netD This user is from outside of this forum
                  divverent@social.vivaldi.netD This user is from outside of this forum
                  divverent@social.vivaldi.net
                  wrote last edited by
                  #38

                  @jonoleth @mjd Pretty sure it's common knowledge that LLMs are nothing but random text generators.

                  OpenAI is a company, not a person. From what I understand, the law banning unlicensed legal advice bans _persons_ and gives them a penalty for doing so anyway.

                  But OpenAI, being a company, cannot commit crimes (after all, how to put a company in prison?). Only the employees can. So the question is which concrete employee committed a crime there. (Yes, some say companies _can_ commit crimes, but then solve the problems by making an employee / owner / ... actually criminally liable - but then they are the ones who have committed the crime)

                  The question is rather, have any employees of OpenAI committed a crime there? If any employee at OpenAI _knew_ that it tries to give legal advice, and did not implement any countermeasures, then that employee has committed a crime. That's the case no matter how the "random text generator" works.

                  If someone tries to get legal advice out of a magic 8-ball, AND the company producing the 8-ball does not implement any countermeasures (such as writing in the manual that responses it gives cannot be used as legal advice), then they can potentially be held liable. Except that in case of a mechanical device that works strikingly like a die it may not be necessary to put such a disclaimer 😉

                  mjd@mathstodon.xyzM jonoleth@mastodon.socialJ 2 Replies Last reply
                  0
                  • sabik@rants.auS sabik@rants.au

                    @divVerent @mjd
                    If ChatGPT misinformed her, that's not very useful

                    divverent@social.vivaldi.netD This user is from outside of this forum
                    divverent@social.vivaldi.netD This user is from outside of this forum
                    divverent@social.vivaldi.net
                    wrote last edited by
                    #39

                    @sabik @mjd As said, that's between her and OpenAI, and does not involve the insurance company.

                    First of all she should be held liable for those vexatious filings.

                    sabik@rants.auS 1 Reply Last reply
                    0
                    • divverent@social.vivaldi.netD divverent@social.vivaldi.net

                      @jonoleth @mjd Pretty sure it's common knowledge that LLMs are nothing but random text generators.

                      OpenAI is a company, not a person. From what I understand, the law banning unlicensed legal advice bans _persons_ and gives them a penalty for doing so anyway.

                      But OpenAI, being a company, cannot commit crimes (after all, how to put a company in prison?). Only the employees can. So the question is which concrete employee committed a crime there. (Yes, some say companies _can_ commit crimes, but then solve the problems by making an employee / owner / ... actually criminally liable - but then they are the ones who have committed the crime)

                      The question is rather, have any employees of OpenAI committed a crime there? If any employee at OpenAI _knew_ that it tries to give legal advice, and did not implement any countermeasures, then that employee has committed a crime. That's the case no matter how the "random text generator" works.

                      If someone tries to get legal advice out of a magic 8-ball, AND the company producing the 8-ball does not implement any countermeasures (such as writing in the manual that responses it gives cannot be used as legal advice), then they can potentially be held liable. Except that in case of a mechanical device that works strikingly like a die it may not be necessary to put such a disclaimer 😉

                      mjd@mathstodon.xyzM This user is from outside of this forum
                      mjd@mathstodon.xyzM This user is from outside of this forum
                      mjd@mathstodon.xyz
                      wrote last edited by
                      #40

                      @divVerent @jonoleth If you're aware of any specific Illinois caselaw that's on point here, I'd be interested to hear about it. But if you're just a nonlawyer making stuff up about what you imagine the law to be, please leave me out of the discussion.

                      divverent@social.vivaldi.netD 1 Reply Last reply
                      0
                      • mjd@mathstodon.xyzM mjd@mathstodon.xyz

                        @divVerent @jonoleth If you're aware of any specific Illinois caselaw that's on point here, I'd be interested to hear about it. But if you're just a nonlawyer making stuff up about what you imagine the law to be, please leave me out of the discussion.

                        divverent@social.vivaldi.netD This user is from outside of this forum
                        divverent@social.vivaldi.netD This user is from outside of this forum
                        divverent@social.vivaldi.net
                        wrote last edited by
                        #41

                        @mjd @jonoleth I am not even American. If in your country machines and companies are "persons" and have human rights that have priority over the human rights of _humans_, then your whole country is wrong. What's next, voting rights in federal elections for corporations? Second Amendment for AIs?

                        But yeah, that might indeed be the case.

                        In my country it is "societas delinquere non potest". A company _cannot_ be defendant of a crime - only the people actually performing the actions can.

                        But yeah, done here. Let's see what broken new case law will come from Trumpistan.

                        teflontrout@beige.partyT 1 Reply Last reply
                        0
                        • divverent@social.vivaldi.netD divverent@social.vivaldi.net

                          @sabik @mjd As said, that's between her and OpenAI, and does not involve the insurance company.

                          First of all she should be held liable for those vexatious filings.

                          sabik@rants.auS This user is from outside of this forum
                          sabik@rants.auS This user is from outside of this forum
                          sabik@rants.au
                          wrote last edited by
                          #42

                          @divVerent @mjd
                          OpenAI is interfering in the insurance company's contract? Or, at least, that's what they're suing for

                          1 Reply Last reply
                          0
                          • divverent@social.vivaldi.netD divverent@social.vivaldi.net

                            @mjd TBH I do not think OpenAI should be responsible. They're just providing a fancy random text generator to the public. And it's outright impossible to teach a random text generator to _not_ output a specific kind of text, as whatever you do, there is a way around it.

                            The woman should pay all costs, as per the usual "vexatious filings" or "frivolous lawsuits" standards.

                            Plus, the law in her state against practicing law without a license starts with "No person shall...". ChatGPT isn't a person.

                            adriano@lile.clA This user is from outside of this forum
                            adriano@lile.clA This user is from outside of this forum
                            adriano@lile.cl
                            wrote last edited by
                            #43

                            @divVerent It's funny how one can use "just" to *just*ify anything. Let's reduce "they hyped their product as «the artificial superintelligence magical clever tool», but didn't even bother adding safety guardrails or disclaimers about output not being legal advice, and now they're getting reamed" as "they're _just_ providing a fancy random text generator to the public."

                            Or let's not.

                            @mjd

                            divverent@social.vivaldi.netD 1 Reply Last reply
                            0
                            • divverent@social.vivaldi.netD divverent@social.vivaldi.net

                              @jonoleth @mjd Pretty sure it's common knowledge that LLMs are nothing but random text generators.

                              OpenAI is a company, not a person. From what I understand, the law banning unlicensed legal advice bans _persons_ and gives them a penalty for doing so anyway.

                              But OpenAI, being a company, cannot commit crimes (after all, how to put a company in prison?). Only the employees can. So the question is which concrete employee committed a crime there. (Yes, some say companies _can_ commit crimes, but then solve the problems by making an employee / owner / ... actually criminally liable - but then they are the ones who have committed the crime)

                              The question is rather, have any employees of OpenAI committed a crime there? If any employee at OpenAI _knew_ that it tries to give legal advice, and did not implement any countermeasures, then that employee has committed a crime. That's the case no matter how the "random text generator" works.

                              If someone tries to get legal advice out of a magic 8-ball, AND the company producing the 8-ball does not implement any countermeasures (such as writing in the manual that responses it gives cannot be used as legal advice), then they can potentially be held liable. Except that in case of a mechanical device that works strikingly like a die it may not be necessary to put such a disclaimer 😉

                              jonoleth@mastodon.socialJ This user is from outside of this forum
                              jonoleth@mastodon.socialJ This user is from outside of this forum
                              jonoleth@mastodon.social
                              wrote last edited by
                              #44

                              @divVerent

                              "Pretty sure it's common knowledge that LLMs are nothing but random text generators."

                              Absolutely not. Maybe in tech circles but the rest of the world has no clue whatsoever how LLMs work. And OpenAI is more than happy to keep it that way

                              "OpenAI is a company, not a person."

                              For legal purposes, most countries treat companies as distinct legal entities, and not just in the US. Still, this is pretty off-topic

                              1 Reply Last reply
                              0
                              • mjd@mathstodon.xyzM mjd@mathstodon.xyz

                                @diazona I don't think it is a shame that this happened to this woman. It appears that she is a very ordinary type of vexatious litigant, except that she is also being aided by ChatGPT.

                                wellsitegeo@masto.aiW This user is from outside of this forum
                                wellsitegeo@masto.aiW This user is from outside of this forum
                                wellsitegeo@masto.ai
                                wrote last edited by
                                #45

                                @mjd @diazona Though I'm not a lawyer (thank Cthulhu, or belly rubs to it's acolyte Menhit @antipope_cats) I do recall the Scottish courts taking exception to a "vexatious litigant" a while ago. It ended badly for said litigant.

                                1 Reply Last reply
                                0
                                • falcennial@mastodon.socialF falcennial@mastodon.social

                                  @mjd @krupo that MIT article demonstrated 5% of AI implementations are profit making. 95% are loss making. so when the investment goldrush mania ends or winds down, 95% of invested amounts will be wiped out. it's billions so it will have a disruptive and negative economic effect that I think we will likely experience as recession. and then that 5% of profit making implementations will be what carries forward, with further investment being modelled on those (anyone can do that right now).

                                  wellsitegeo@masto.aiW This user is from outside of this forum
                                  wellsitegeo@masto.aiW This user is from outside of this forum
                                  wellsitegeo@masto.ai
                                  wrote last edited by
                                  #46

                                  @falcennial @mjd @krupo
                                  Looking to the foreseeable future, the AI boosters (especially those using AI to write their "opinions") will blame the bubble-burst on the Trump-Bibi War (no, they can't escape the blame) and *it's* crash.
                                  How much are Darien Scheme share certificates worth these days?

                                  1 Reply Last reply
                                  0
                                  • marshray@infosec.exchangeM marshray@infosec.exchange

                                    @mjd “41. On October 29, 2025, OPENAI amended the terms and usage policies of ChatGPT to prohibit users from using ChatGPT to provide tailored legal advice. Prior to the October 29, 2025 emendation, ChatGPT’s terms of use did not prohibit users from using ChatGPT to draft legal papers, conduct legal research, provide legal analysis or give legal advice.”

                                    wellsitegeo@masto.aiW This user is from outside of this forum
                                    wellsitegeo@masto.aiW This user is from outside of this forum
                                    wellsitegeo@masto.ai
                                    wrote last edited by
                                    #47

                                    @marshray @mjd
                                    Any dates for the (alleged, legal modesty board) AI legal advice?

                                    What's that, Lassie? You hear the sound of distant hard-drive shredders working overtime?

                                    marshray@infosec.exchangeM 1 Reply Last reply
                                    0
                                    • divverent@social.vivaldi.netD divverent@social.vivaldi.net

                                      @mjd TBH I do not think OpenAI should be responsible. They're just providing a fancy random text generator to the public. And it's outright impossible to teach a random text generator to _not_ output a specific kind of text, as whatever you do, there is a way around it.

                                      The woman should pay all costs, as per the usual "vexatious filings" or "frivolous lawsuits" standards.

                                      Plus, the law in her state against practicing law without a license starts with "No person shall...". ChatGPT isn't a person.

                                      wellsitegeo@masto.aiW This user is from outside of this forum
                                      wellsitegeo@masto.aiW This user is from outside of this forum
                                      wellsitegeo@masto.ai
                                      wrote last edited by
                                      #48

                                      @divVerent @mjd "barratry" - that's a term I remember from the Scottish Legal High Heidjuns dealing with a similar "vexatious litigant". I remember looking it up (but not the precise definition).

                                      1 Reply Last reply
                                      0
                                      • jonoleth@mastodon.socialJ jonoleth@mastodon.social

                                        @divVerent @mjd ChatGPT is not a person, which is why ChatGPT is not being sued. OpenAI sells a tool that gave her legal advice, and they certainly didn't say anywhere that it's actually just a "fancy random text generator"

                                        wellsitegeo@masto.aiW This user is from outside of this forum
                                        wellsitegeo@masto.aiW This user is from outside of this forum
                                        wellsitegeo@masto.ai
                                        wrote last edited by
                                        #49

                                        @jonoleth @divVerent @mjd

                                        Wait, what?
                                        They *sell* this shit? And charge money for it?

                                        Where the holy cat turds do they find clients? On the Internet?

                                        (No, I've never tried to use an AI.)

                                        1 Reply Last reply
                                        0
                                        • adriano@lile.clA adriano@lile.cl

                                          @divVerent It's funny how one can use "just" to *just*ify anything. Let's reduce "they hyped their product as «the artificial superintelligence magical clever tool», but didn't even bother adding safety guardrails or disclaimers about output not being legal advice, and now they're getting reamed" as "they're _just_ providing a fancy random text generator to the public."

                                          Or let's not.

                                          @mjd

                                          divverent@social.vivaldi.netD This user is from outside of this forum
                                          divverent@social.vivaldi.netD This user is from outside of this forum
                                          divverent@social.vivaldi.net
                                          wrote last edited by
                                          #50

                                          @adriano @mjd There already is a disclaimer:

                                          Link Preview Image
                                          Terms of use

                                          favicon

                                          (web.archive.org)

                                          > What you cannot do. You may not use our Services for any illegal, harmful, or abusive activity. For example, you may not:
                                          [...]
                                          > Represent that Output was human-generated when it was not.
                                          [...]
                                          > Output may not always be accurate. You should not rely on Output from our Services as a sole source of truth or factual information, or as a substitute for professional advice.
                                          [...]
                                          > You must not use any Output relating to a person for any purpose that could have a legal or material impact on that person, such as making credit, educational, employment, housing, insurance, legal, medical, or other important decisions about them.

                                          Pretty much looks like clearly disclaiming to me.

                                          adriano@lile.clA 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups