Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. A new twist in the "AI license laundering of chardet" story https://github.com/chardet/chardet/issues/327

A new twist in the "AI license laundering of chardet" story https://github.com/chardet/chardet/issues/327

Scheduled Pinned Locked Moved Uncategorized
52 Posts 38 Posters 135 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • cwebber@social.coopC cwebber@social.coop

    A new twist in the "AI license laundering of chardet" story https://github.com/chardet/chardet/issues/327

    ralph_social@dresden.networkR This user is from outside of this forum
    ralph_social@dresden.networkR This user is from outside of this forum
    ralph_social@dresden.network
    wrote last edited by
    #42

    Krass, dass sich AI-Firmen einfach Open Source Code schnappen und die Lizenzen "waschen" wollen. 😤

    Das ist genau das Problem mit dem aktuellen AI-Hype: Die großen Player denken, sie können einfach alles verwenden was im Netz steht. Und wenn's rechtlich eng wird, wird halt schnell die Lizenz geändert...

    Respekt an Mark Pilgrim dass er sich dagegen wehrt! Open Source lebt von Vertrauen und klaren Regeln - nicht von solchen Manövern.

    #OpenSource #AIEthics #Licensing

    1 Reply Last reply
    0
    • cwebber@social.coopC cwebber@social.coop

      A new twist in the "AI license laundering of chardet" story https://github.com/chardet/chardet/issues/327

      G This user is from outside of this forum
      G This user is from outside of this forum
      gerardthornley@hachyderm.io
      wrote last edited by
      #43

      @cwebber Reading through all the comments there left me wondering if anyone has (yet) hooked up an LLM to be a project maintainer. Interactions via issues and just let it loose. People would be utterly mad to ever include it in their supply chain, and yet people do do mad things.

      1 Reply Last reply
      0
      • cwebber@social.coopC cwebber@social.coop

        A new twist in the "AI license laundering of chardet" story https://github.com/chardet/chardet/issues/327

        avirr@sfba.socialA This user is from outside of this forum
        avirr@sfba.socialA This user is from outside of this forum
        avirr@sfba.social
        wrote last edited by
        #44

        @cwebber Isn’t this what forks are for?

        1 Reply Last reply
        0
        • feld@friedcheese.usF feld@friedcheese.us
          @cwebber

          > Their claim that it is a "complete rewrite" is irrelevant, since they had ample exposure to the originally licensed code (i.e. this is not a "clean room" implementation). Adding a fancy code generator into the mix does not somehow grant them any additional rights.

          The human didn't write the code, the LLM did. "They" which had "ample exposure to the originally licensed code" does not exist; "they" are ephemeral.

          1. Start a fresh session / clean context, make it meticulously document the architecture, APIs, etc

          2. keep those documents, throw away the code, start a new session with an LLM that has clean context and tell it to build off those documents.

          That's clean room. If the original code was not in the LLM's context, it's not violating the license.

          This is how you can do this. Proving beyond a reasonable doubt he didn't do it this way is going to require a lot of evidence nobody will have.
          vv@solarpunk.moeV This user is from outside of this forum
          vv@solarpunk.moeV This user is from outside of this forum
          vv@solarpunk.moe
          wrote last edited by
          #45

          @feld @cwebber the AI is still trained on the code beforehand

          vv@solarpunk.moeV 1 Reply Last reply
          0
          • vv@solarpunk.moeV vv@solarpunk.moe

            @feld @cwebber the AI is still trained on the code beforehand

            vv@solarpunk.moeV This user is from outside of this forum
            vv@solarpunk.moeV This user is from outside of this forum
            vv@solarpunk.moe
            wrote last edited by
            #46

            @feld @cwebber a "clean context" doesn't mean that there's no training data, it's still trained on a bunch of source code which likely includes the original

            feld@friedcheese.usF 1 Reply Last reply
            0
            • vv@solarpunk.moeV vv@solarpunk.moe

              @feld @cwebber a "clean context" doesn't mean that there's no training data, it's still trained on a bunch of source code which likely includes the original

              feld@friedcheese.usF This user is from outside of this forum
              feld@friedcheese.usF This user is from outside of this forum
              feld@friedcheese.us
              wrote last edited by
              #47
              @vv @cwebber proving the original was trained by the model or is in the model is quite difficult to do and is questionable whether or not it really matters anyway.

              Chris Lattner was "trained on" GCC when he wrote LLVM. He studied it a lot. GCC compiles code C/C++ successfully, LLVM compiles C/C++ code successfully.

              Both produce completely working bytecode and generally you don't *need* one compiler over the other to get an end result that is acceptable.

              Should LLVM be allowed to have an Apache license because of this?

              These are tough questions.
              1 Reply Last reply
              0
              • cwebber@social.coopC cwebber@social.coop

                omg I am just seeing now that the dude who did the "AI relicensing" fucking replied with an obvious slop response, of all the fucking disrespectful things to do, holy fucking shit https://github.com/chardet/chardet/issues/327#issuecomment-4005195078

                pikesley@mastodon.me.ukP This user is from outside of this forum
                pikesley@mastodon.me.ukP This user is from outside of this forum
                pikesley@mastodon.me.uk
                wrote last edited by
                #48

                @cwebber I felt my brain getting smoother as I read that

                1 Reply Last reply
                0
                • cwebber@social.coopC cwebber@social.coop

                  Winning option 1: yes, you can vibe code proprietary codebases into the public domain, allowing us to bootstrap proprietary codebases quickly

                  Winning option 2: stopping laundering of copyleft codebases

                  Either of these are interesting outcomes!

                  svines@gts.svines.rodeoS This user is from outside of this forum
                  svines@gts.svines.rodeoS This user is from outside of this forum
                  svines@gts.svines.rodeo
                  wrote last edited by
                  #49

                  @cwebber Microslop committed to picking up the legal bill for anyone concerned about copyright issues with AI outputs from copilot so one could hypothetically use their tools to "clean room" implement Photoshop and then have Satya fight Adobe for your right to do so. Sounds fun to me!

                  https://blogs.microsoft.com/on-the-issues/2023/09/07/copilot-copyright-commitment-ai-legal-concerns/

                  valpackett@social.treehouse.systemsV borisbarbour@mastodon.socialB 2 Replies Last reply
                  0
                  • svines@gts.svines.rodeoS svines@gts.svines.rodeo

                    @cwebber Microslop committed to picking up the legal bill for anyone concerned about copyright issues with AI outputs from copilot so one could hypothetically use their tools to "clean room" implement Photoshop and then have Satya fight Adobe for your right to do so. Sounds fun to me!

                    https://blogs.microsoft.com/on-the-issues/2023/09/07/copilot-copyright-commitment-ai-legal-concerns/

                    valpackett@social.treehouse.systemsV This user is from outside of this forum
                    valpackett@social.treehouse.systemsV This user is from outside of this forum
                    valpackett@social.treehouse.systems
                    wrote last edited by
                    #50

                    @svines @cwebber call that Project Photoslop

                    1 Reply Last reply
                    0
                    • svines@gts.svines.rodeoS svines@gts.svines.rodeo

                      @cwebber Microslop committed to picking up the legal bill for anyone concerned about copyright issues with AI outputs from copilot so one could hypothetically use their tools to "clean room" implement Photoshop and then have Satya fight Adobe for your right to do so. Sounds fun to me!

                      https://blogs.microsoft.com/on-the-issues/2023/09/07/copilot-copyright-commitment-ai-legal-concerns/

                      borisbarbour@mastodon.socialB This user is from outside of this forum
                      borisbarbour@mastodon.socialB This user is from outside of this forum
                      borisbarbour@mastodon.social
                      wrote last edited by
                      #51

                      @svines @cwebber

                      This is topical:

                      "To protect against this, customers ... must not attempt to generate infringing materials, including not providing input to a Copilot service that the customer does not have appropriate rights to use."

                      1 Reply Last reply
                      0
                      • cstanhope@social.coopC cstanhope@social.coop

                        @cwebber I'm not sure that's slop, but I won't discount the possibility... 🤔 But this part is funny in the dark humor sort of way:

                        "...explicitly instructed Claude not to base anything on LGPL/GPL-licensed code."

                        So, you see, no problem... 🙄

                        lukeharby@infosec.exchangeL This user is from outside of this forum
                        lukeharby@infosec.exchangeL This user is from outside of this forum
                        lukeharby@infosec.exchange
                        wrote last edited by
                        #52

                        @cstanhope @cwebber

                        Claude after being explicitly instructed not to base anything on LGPL/GPL-licensed code

                        Link Preview Image
                        1 Reply Last reply
                        0
                        • R relay@relay.infosec.exchange shared this topic
                        Reply
                        • Reply as topic
                        Log in to reply
                        • Oldest to Newest
                        • Newest to Oldest
                        • Most Votes


                        • Login

                        • Login or register to search.
                        • First post
                          Last post
                        0
                        • Categories
                        • Recent
                        • Tags
                        • Popular
                        • World
                        • Users
                        • Groups