Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Cyborg)
  • No Skin
Collapse
Brand Logo

CIRCLE WITH A DOT

  1. Home
  2. Uncategorized
  3. Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool.

Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool.

Scheduled Pinned Locked Moved Uncategorized
39 Posts 20 Posters 8 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • nirak@carhenge.clubN nirak@carhenge.club

    @screwturn @briankrebs What if the summaries are wrong? How do you know? If you read through everything to find the errors, does it actually save time?

    screwturn@mastodon.socialS This user is from outside of this forum
    screwturn@mastodon.socialS This user is from outside of this forum
    screwturn@mastodon.social
    wrote last edited by
    #17

    @nirak

    Wrong in what way?
    Yes, in most cases I'm reading the entire text, but sometimes the AI captures something I missed, and other times it confirms what I already got.

    Time saving does feature, but the bigger issue is that using it improves validity, because of catching the missed topics

    @briankrebs

    stumpythemutt@social.linux.pizzaS briankrebs@infosec.exchangeB 2 Replies Last reply
    0
    • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

      @nirak @screwturn Spot on. If you have to redo someone else's work all the time because you're not sure if it's right, why not just do that work yourself from the get-go?

      screwturn@mastodon.socialS This user is from outside of this forum
      screwturn@mastodon.socialS This user is from outside of this forum
      screwturn@mastodon.social
      wrote last edited by
      #18

      @briankrebs

      In qualitative research we routinely redo each other's work and our own.
      Having an AI do that too increases construction validity and reliability.

      @nirak

      1 Reply Last reply
      0
      • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

        Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool. What a mess. And a tough one to bounce back from. I get asked all the time how I use AI in my work, and my answer is always the same: I don't, for all the reasons I also don't delegate important research to others, plus a whole bunch of other good reasons. But I really am interested in the answer from other journalists, because I suspect I'm in the minority here.

        From Futurism.com:

        "In the post, Edwards said that he was sick at the time, and “while working from bed with a fever and very little sleep,” he “unintentionally made a serious journalistic error” as he attempted to use an “experimental Claude Code-based AI tool” to help him “extract relevant verbatim source material.” He said the tool wasn’t being used to generate the article, but was instead designed to “help list structured references” to put in an outline. When the tool failed to work, said Edwards, he decided to try and use ChatGPT to help him understand why.

        “I should have taken a sick day because in the course of that interaction, I inadvertently ended up with a paraphrased version of Shambaugh’s words rather than his actual words,” Edwards continued. He emphasized that the “text of the article was human-written by us, and this incident was isolated and is not representative of Ars‘ editorial standards. None of our articles are AI-generated, it is against company policy and we have always respected that.”

        Link Preview Image
        Ars Technica Fires Reporter After AI Controversy Involving Fabricated Quotes

        Ars Technica has fired senior AI reporter Benj Edwards following an outrage-sparking controversy involving AI-fabricated quotes.

        favicon

        Futurism (futurism.com)

        chicob@mstdn.socialC This user is from outside of this forum
        chicob@mstdn.socialC This user is from outside of this forum
        chicob@mstdn.social
        wrote last edited by
        #19

        @briankrebs
        AI in journalism is Farse Technica

        briankrebs@infosec.exchangeB 1 Reply Last reply
        0
        • chicob@mstdn.socialC chicob@mstdn.social

          @briankrebs
          AI in journalism is Farse Technica

          briankrebs@infosec.exchangeB This user is from outside of this forum
          briankrebs@infosec.exchangeB This user is from outside of this forum
          briankrebs@infosec.exchange
          wrote last edited by
          #20

          @chicob Arse. It was right there, dude.

          1 Reply Last reply
          0
          • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

            Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool. What a mess. And a tough one to bounce back from. I get asked all the time how I use AI in my work, and my answer is always the same: I don't, for all the reasons I also don't delegate important research to others, plus a whole bunch of other good reasons. But I really am interested in the answer from other journalists, because I suspect I'm in the minority here.

            From Futurism.com:

            "In the post, Edwards said that he was sick at the time, and “while working from bed with a fever and very little sleep,” he “unintentionally made a serious journalistic error” as he attempted to use an “experimental Claude Code-based AI tool” to help him “extract relevant verbatim source material.” He said the tool wasn’t being used to generate the article, but was instead designed to “help list structured references” to put in an outline. When the tool failed to work, said Edwards, he decided to try and use ChatGPT to help him understand why.

            “I should have taken a sick day because in the course of that interaction, I inadvertently ended up with a paraphrased version of Shambaugh’s words rather than his actual words,” Edwards continued. He emphasized that the “text of the article was human-written by us, and this incident was isolated and is not representative of Ars‘ editorial standards. None of our articles are AI-generated, it is against company policy and we have always respected that.”

            Link Preview Image
            Ars Technica Fires Reporter After AI Controversy Involving Fabricated Quotes

            Ars Technica has fired senior AI reporter Benj Edwards following an outrage-sparking controversy involving AI-fabricated quotes.

            favicon

            Futurism (futurism.com)

            colo_lee@mstdn.socialC This user is from outside of this forum
            colo_lee@mstdn.socialC This user is from outside of this forum
            colo_lee@mstdn.social
            wrote last edited by
            #21

            @briankrebs well, at least he only did it once.
            er, correction ...
            he only got *caught* doing it once

            1 Reply Last reply
            0
            • screwturn@mastodon.socialS screwturn@mastodon.social

              @nirak

              Wrong in what way?
              Yes, in most cases I'm reading the entire text, but sometimes the AI captures something I missed, and other times it confirms what I already got.

              Time saving does feature, but the bigger issue is that using it improves validity, because of catching the missed topics

              @briankrebs

              stumpythemutt@social.linux.pizzaS This user is from outside of this forum
              stumpythemutt@social.linux.pizzaS This user is from outside of this forum
              stumpythemutt@social.linux.pizza
              wrote last edited by
              #22

              @screwturn @nirak @briankrebs Even a blind pig will find the occasional acorn.

              screwturn@mastodon.socialS 1 Reply Last reply
              0
              • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool. What a mess. And a tough one to bounce back from. I get asked all the time how I use AI in my work, and my answer is always the same: I don't, for all the reasons I also don't delegate important research to others, plus a whole bunch of other good reasons. But I really am interested in the answer from other journalists, because I suspect I'm in the minority here.

                From Futurism.com:

                "In the post, Edwards said that he was sick at the time, and “while working from bed with a fever and very little sleep,” he “unintentionally made a serious journalistic error” as he attempted to use an “experimental Claude Code-based AI tool” to help him “extract relevant verbatim source material.” He said the tool wasn’t being used to generate the article, but was instead designed to “help list structured references” to put in an outline. When the tool failed to work, said Edwards, he decided to try and use ChatGPT to help him understand why.

                “I should have taken a sick day because in the course of that interaction, I inadvertently ended up with a paraphrased version of Shambaugh’s words rather than his actual words,” Edwards continued. He emphasized that the “text of the article was human-written by us, and this incident was isolated and is not representative of Ars‘ editorial standards. None of our articles are AI-generated, it is against company policy and we have always respected that.”

                Link Preview Image
                Ars Technica Fires Reporter After AI Controversy Involving Fabricated Quotes

                Ars Technica has fired senior AI reporter Benj Edwards following an outrage-sparking controversy involving AI-fabricated quotes.

                favicon

                Futurism (futurism.com)

                kennethbousquet@mastodon.socialK This user is from outside of this forum
                kennethbousquet@mastodon.socialK This user is from outside of this forum
                kennethbousquet@mastodon.social
                wrote last edited by
                #23

                @briankrebs He's not at fault here. Even after explanations, he souldn't have been fired. The person has not intentionaly take credit for others work. That person should be reinstated, have his job back.

                1 Reply Last reply
                0
                • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                  Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool. What a mess. And a tough one to bounce back from. I get asked all the time how I use AI in my work, and my answer is always the same: I don't, for all the reasons I also don't delegate important research to others, plus a whole bunch of other good reasons. But I really am interested in the answer from other journalists, because I suspect I'm in the minority here.

                  From Futurism.com:

                  "In the post, Edwards said that he was sick at the time, and “while working from bed with a fever and very little sleep,” he “unintentionally made a serious journalistic error” as he attempted to use an “experimental Claude Code-based AI tool” to help him “extract relevant verbatim source material.” He said the tool wasn’t being used to generate the article, but was instead designed to “help list structured references” to put in an outline. When the tool failed to work, said Edwards, he decided to try and use ChatGPT to help him understand why.

                  “I should have taken a sick day because in the course of that interaction, I inadvertently ended up with a paraphrased version of Shambaugh’s words rather than his actual words,” Edwards continued. He emphasized that the “text of the article was human-written by us, and this incident was isolated and is not representative of Ars‘ editorial standards. None of our articles are AI-generated, it is against company policy and we have always respected that.”

                  Link Preview Image
                  Ars Technica Fires Reporter After AI Controversy Involving Fabricated Quotes

                  Ars Technica has fired senior AI reporter Benj Edwards following an outrage-sparking controversy involving AI-fabricated quotes.

                  favicon

                  Futurism (futurism.com)

                  reflex@retrogaming.socialR This user is from outside of this forum
                  reflex@retrogaming.socialR This user is from outside of this forum
                  reflex@retrogaming.social
                  wrote last edited by
                  #24

                  @briankrebs I learned not to trust Ars reporting after the Hacker X story, which they have still declined to retract.

                  1 Reply Last reply
                  0
                  • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                    Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool. What a mess. And a tough one to bounce back from. I get asked all the time how I use AI in my work, and my answer is always the same: I don't, for all the reasons I also don't delegate important research to others, plus a whole bunch of other good reasons. But I really am interested in the answer from other journalists, because I suspect I'm in the minority here.

                    From Futurism.com:

                    "In the post, Edwards said that he was sick at the time, and “while working from bed with a fever and very little sleep,” he “unintentionally made a serious journalistic error” as he attempted to use an “experimental Claude Code-based AI tool” to help him “extract relevant verbatim source material.” He said the tool wasn’t being used to generate the article, but was instead designed to “help list structured references” to put in an outline. When the tool failed to work, said Edwards, he decided to try and use ChatGPT to help him understand why.

                    “I should have taken a sick day because in the course of that interaction, I inadvertently ended up with a paraphrased version of Shambaugh’s words rather than his actual words,” Edwards continued. He emphasized that the “text of the article was human-written by us, and this incident was isolated and is not representative of Ars‘ editorial standards. None of our articles are AI-generated, it is against company policy and we have always respected that.”

                    Link Preview Image
                    Ars Technica Fires Reporter After AI Controversy Involving Fabricated Quotes

                    Ars Technica has fired senior AI reporter Benj Edwards following an outrage-sparking controversy involving AI-fabricated quotes.

                    favicon

                    Futurism (futurism.com)

                    teriradichel@infosec.exchangeT This user is from outside of this forum
                    teriradichel@infosec.exchangeT This user is from outside of this forum
                    teriradichel@infosec.exchange
                    wrote last edited by
                    #25

                    @briankrebs you 100% cannot trust it. Like Google search results and Wikipedia. But it still might give you some idea or thought or resource you hadn’t seen yet that you can go research further. It can help you think of new questions and point you in new directions (which can be good or bad). I use it to explore ideas and if I do copy something written by AI I write “from Google AI:” or whatever so people can take it with a grain of salt and can back that up with links to other sources. It’s usually something I know is right but I like the way it wrote it and saves me some time. Sometimes I call out when it is wrong to demonstrate why you can’t always trust it. But I’m researching and writing about AI not the kind of things you write about so it’s a bit different. I generally just cite sources if I’m writing something about a data breach like you (and nowhere near the deep dive you do!)

                    1 Reply Last reply
                    0
                    • screwturn@mastodon.socialS screwturn@mastodon.social

                      @nirak

                      Wrong in what way?
                      Yes, in most cases I'm reading the entire text, but sometimes the AI captures something I missed, and other times it confirms what I already got.

                      Time saving does feature, but the bigger issue is that using it improves validity, because of catching the missed topics

                      @briankrebs

                      briankrebs@infosec.exchangeB This user is from outside of this forum
                      briankrebs@infosec.exchangeB This user is from outside of this forum
                      briankrebs@infosec.exchange
                      wrote last edited by
                      #26

                      @screwturn @nirak There are some pretty decent and recent studies showing AI substantially misses or misrepresents the point or summary of a story about 40-50 percent of the time.

                      ct@app.wafrn.netC screwturn@mastodon.socialS 2 Replies Last reply
                      0
                      • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                        @screwturn @nirak There are some pretty decent and recent studies showing AI substantially misses or misrepresents the point or summary of a story about 40-50 percent of the time.

                        ct@app.wafrn.netC This user is from outside of this forum
                        ct@app.wafrn.netC This user is from outside of this forum
                        ct@app.wafrn.net
                        wrote last edited by
                        #27

                        Even if the success were 95%, as a journalist, consistently using a stochastic method to give sources guarantees you eventually fuck up and let a fabricated quote into print.

                        screwturn@mastodon.socialS 1 Reply Last reply
                        0
                        • ct@app.wafrn.netC ct@app.wafrn.net

                          Even if the success were 95%, as a journalist, consistently using a stochastic method to give sources guarantees you eventually fuck up and let a fabricated quote into print.

                          screwturn@mastodon.socialS This user is from outside of this forum
                          screwturn@mastodon.socialS This user is from outside of this forum
                          screwturn@mastodon.social
                          wrote last edited by
                          #28

                          @ct
                          um... sure, but who is going to be asking the free version of ChatGPT for sources?
                          That is going to be a very poor use case.

                          If I am using the Ai that is inside my CAQDAS, I am not going to see hallucination, and it internally cites each fact it produces. Reliability and validity are going to vary greatly depending on the environment you use the AI in, and what you are trying to do.

                          @briankrebs @nirak

                          ct@app.wafrn.netC 1 Reply Last reply
                          0
                          • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                            @screwturn @nirak There are some pretty decent and recent studies showing AI substantially misses or misrepresents the point or summary of a story about 40-50 percent of the time.

                            screwturn@mastodon.socialS This user is from outside of this forum
                            screwturn@mastodon.socialS This user is from outside of this forum
                            screwturn@mastodon.social
                            wrote last edited by
                            #29

                            @briankrebs
                            I'm not sure what that means
                            Like 50% of the time I use it, it will miss at least one point? Sure, but those odds are fine by me if it is also spotting things I missed, and has a reasonable inter-rater reliability with what I saw.

                            If you mean it gets 50% of the points wrong, then that is probably true of the free versions, but not what I am seeing in practice when I use the AI inside my CAQDAS

                            @nirak

                            1 Reply Last reply
                            0
                            • stumpythemutt@social.linux.pizzaS stumpythemutt@social.linux.pizza

                              @screwturn @nirak @briankrebs Even a blind pig will find the occasional acorn.

                              screwturn@mastodon.socialS This user is from outside of this forum
                              screwturn@mastodon.socialS This user is from outside of this forum
                              screwturn@mastodon.social
                              wrote last edited by
                              #30

                              @StumpyTheMutt
                              If it finds an acorn that I missed, then it found something of value.

                              Keep in mind, I'm not using the free version in its wide-open configuration, but rather a tightly configured version inside a research workbench. In three years of use, I have not seen a single case of hallucination by the LLM.

                              @nirak @briankrebs

                              1 Reply Last reply
                              0
                              • R relay@relay.an.exchange shared this topic
                              • screwturn@mastodon.socialS screwturn@mastodon.social

                                @ct
                                um... sure, but who is going to be asking the free version of ChatGPT for sources?
                                That is going to be a very poor use case.

                                If I am using the Ai that is inside my CAQDAS, I am not going to see hallucination, and it internally cites each fact it produces. Reliability and validity are going to vary greatly depending on the environment you use the AI in, and what you are trying to do.

                                @briankrebs @nirak

                                ct@app.wafrn.netC This user is from outside of this forum
                                ct@app.wafrn.netC This user is from outside of this forum
                                ct@app.wafrn.net
                                wrote last edited by
                                #31

                                I'm curious how your example actually works under the hood.

                                I have a sneaking suspicion that maybe your personal experience with a research summarization tool was not relevant to this story of a tech journalist, who needs to source current events from myriad sources and not just a limited database of pre-curated published research? I speculate your CAQDAS tool would not have been useful for a current events journalist who may need to quote things like statements from leadership, self-published cybersecurity reports, transcriptions of tech presentations etc… where there's a lot more critical thinking involved in selecting who to source from.

                                Regardless, I'd love to see how your CAQDAS tool fares against peer-reviewed fact-checking tests. I am very skeptical the failure rate is under 1% just from your testimony.

                                screwturn@mastodon.socialS 1 Reply Last reply
                                0
                                • ct@app.wafrn.netC ct@app.wafrn.net

                                  I'm curious how your example actually works under the hood.

                                  I have a sneaking suspicion that maybe your personal experience with a research summarization tool was not relevant to this story of a tech journalist, who needs to source current events from myriad sources and not just a limited database of pre-curated published research? I speculate your CAQDAS tool would not have been useful for a current events journalist who may need to quote things like statements from leadership, self-published cybersecurity reports, transcriptions of tech presentations etc… where there's a lot more critical thinking involved in selecting who to source from.

                                  Regardless, I'd love to see how your CAQDAS tool fares against peer-reviewed fact-checking tests. I am very skeptical the failure rate is under 1% just from your testimony.

                                  screwturn@mastodon.socialS This user is from outside of this forum
                                  screwturn@mastodon.socialS This user is from outside of this forum
                                  screwturn@mastodon.social
                                  wrote last edited by
                                  #32

                                  @ct

                                  I have a sneaking suspicion that you don't know what "goes on under the hood" of qualitative research.
                                  Let's just be clear what we are talking about in praxis.

                                  One would not use the AI to find material, conduct interviews (although that is a distinct future possibility), or to do the discovery part of research, or journalism.
                                  Once you have pulled those texts, transcripts, etc into a CAQDAS, THEN you would use the AI to summarize, identify topics, match topics, etc

                                  @briankrebs @nirak

                                  screwturn@mastodon.socialS ct@app.wafrn.netC 2 Replies Last reply
                                  0
                                  • screwturn@mastodon.socialS screwturn@mastodon.social

                                    @ct

                                    I have a sneaking suspicion that you don't know what "goes on under the hood" of qualitative research.
                                    Let's just be clear what we are talking about in praxis.

                                    One would not use the AI to find material, conduct interviews (although that is a distinct future possibility), or to do the discovery part of research, or journalism.
                                    Once you have pulled those texts, transcripts, etc into a CAQDAS, THEN you would use the AI to summarize, identify topics, match topics, etc

                                    @briankrebs @nirak

                                    screwturn@mastodon.socialS This user is from outside of this forum
                                    screwturn@mastodon.socialS This user is from outside of this forum
                                    screwturn@mastodon.social
                                    wrote last edited by
                                    #33

                                    @ct
                                    "peer-reviewed fact-checking tests"

                                    Do you mean inter-rater reliability?
                                    Because several people have found various LLMs in CAQDAS platforms to be on par between the AI and human researchers and between human researchers

                                    @briankrebs @nirak

                                    1 Reply Last reply
                                    0
                                    • michael@westergaard.socialM michael@westergaard.social
                                      "We always write things by hand and never use AI, except for this one small case where you caught us. And the next time you catch us. But there's no general tendency. You're just very good at catching exactly the cases where we use AI."
                                      kasperd@westergaard.socialK This user is from outside of this forum
                                      kasperd@westergaard.socialK This user is from outside of this forum
                                      kasperd@westergaard.social
                                      wrote last edited by
                                      #34

                                      I would expect that journalists have to deal with information from untrustworthy sources all the time. In that regard the output of an AI might not be worse than a lot of the other misinformation they are juggling.

                                      Using AI at some point during the process is not guaranteed to result in a worse end result if the journalist is otherwise doing a good job.

                                      Of course it's possible for a journalist to do a bad job such as including AI output verbatim in the final product without validating the correctness. But bad journalism isn't a novel concept. There has been journalists producing bad results before AI.

                                      michael@westergaard.socialM 1 Reply Last reply
                                      0
                                      • briankrebs@infosec.exchangeB briankrebs@infosec.exchange

                                        Found myself wincing while reading this story about how Ars Technica fired a reporter over fabricated quotations generated by an AI tool. What a mess. And a tough one to bounce back from. I get asked all the time how I use AI in my work, and my answer is always the same: I don't, for all the reasons I also don't delegate important research to others, plus a whole bunch of other good reasons. But I really am interested in the answer from other journalists, because I suspect I'm in the minority here.

                                        From Futurism.com:

                                        "In the post, Edwards said that he was sick at the time, and “while working from bed with a fever and very little sleep,” he “unintentionally made a serious journalistic error” as he attempted to use an “experimental Claude Code-based AI tool” to help him “extract relevant verbatim source material.” He said the tool wasn’t being used to generate the article, but was instead designed to “help list structured references” to put in an outline. When the tool failed to work, said Edwards, he decided to try and use ChatGPT to help him understand why.

                                        “I should have taken a sick day because in the course of that interaction, I inadvertently ended up with a paraphrased version of Shambaugh’s words rather than his actual words,” Edwards continued. He emphasized that the “text of the article was human-written by us, and this incident was isolated and is not representative of Ars‘ editorial standards. None of our articles are AI-generated, it is against company policy and we have always respected that.”

                                        Link Preview Image
                                        Ars Technica Fires Reporter After AI Controversy Involving Fabricated Quotes

                                        Ars Technica has fired senior AI reporter Benj Edwards following an outrage-sparking controversy involving AI-fabricated quotes.

                                        favicon

                                        Futurism (futurism.com)

                                        mlanger@mastodon.worldM This user is from outside of this forum
                                        mlanger@mastodon.worldM This user is from outside of this forum
                                        mlanger@mastodon.world
                                        wrote last edited by
                                        #35

                                        @briankrebs @rpmik I think using AI to do your work is like playing Russian Roulette.

                                        1 Reply Last reply
                                        0
                                        • screwturn@mastodon.socialS screwturn@mastodon.social

                                          @ct

                                          I have a sneaking suspicion that you don't know what "goes on under the hood" of qualitative research.
                                          Let's just be clear what we are talking about in praxis.

                                          One would not use the AI to find material, conduct interviews (although that is a distinct future possibility), or to do the discovery part of research, or journalism.
                                          Once you have pulled those texts, transcripts, etc into a CAQDAS, THEN you would use the AI to summarize, identify topics, match topics, etc

                                          @briankrebs @nirak

                                          ct@app.wafrn.netC This user is from outside of this forum
                                          ct@app.wafrn.netC This user is from outside of this forum
                                          ct@app.wafrn.net
                                          wrote last edited by
                                          #36

                                          I don't… that's why I said I was curious what it was. I was asking.
                                          Anyway… if your tool can't be used for the task the article is about, which is tech journalism. why did you come in to the article's replies to defend AI research? All I'm seeing is you reacting to criticism of an example where LLM for research didn't work by defending your own use, in an application and implementation that both aren't relevant to the event at hand.

                                          screwturn@mastodon.socialS 1 Reply Last reply
                                          0
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • World
                                          • Users
                                          • Groups