• @Aurenkin@sh.itjust.works
    link
    fedilink
    8
    edit-2
    1 year ago

    Smells like bullshit. The graphs they showed in the source paper with their accuracy at like 100% for every test seem even more like bullshit. Did they run the model over the training data or what?

    Maybe I’m wrong but text is just way too high signal to noise medium to be able to tell if it was written by an AI. The false positives would be high enough that it’s effectively useless. Does anyone have another perspective on this? If I’m missing some nuance here I’d love to understand more.

    • @Shdwdrgn@mander.xyz
      link
      fedilink
      English
      41 year ago

      I’m just guessing here, but if you have a way to train your model with the same or similar types of data as the model you are comparing with, then you have a high probability of detecting when the output is similar. To put it another way, if I ask chatGPT its favorite red condiment and it answers “mustard”, then if my model is trained to give the same answer it makes it easy to detect something that was likely created by an AI. If the answer doesn’t match what my model spits out, then it has a higher chance of being human-generated. Does that make sense?

      • @Aurenkin@sh.itjust.works
        link
        fedilink
        31 year ago

        Yeah that makes sense. I’m still very sceptical though because as your example illustrates, it’s perfectly valid for a human to answer “mustard” as well, plus there is an element of randomness inserted into the model output. Maybe it’s doable but I’m unconvinced that you can meaningfully distinguish between human and AI written text. Unless you make a detector that looks for “As a large language model…” Then maybe it can detect ChatGPT specifically.

        • @Shdwdrgn@mander.xyz
          link
          fedilink
          English
          21 year ago

          Agreed, even a perfectly trained clone of chatGPT wouldn’t get that high a hit rate, although I do think that the larger the article being compared, the better its chances would be of making an accurate prediction. The thing is that we soon won’t actually be able to tell the difference as computers get smarter. Sees like right now the only practical application is for kids to cheat on their homework, but what happens when it gets smart enough to write actual research papers with unique proofs?

          • @Hanabie@sh.itjust.works
            link
            fedilink
            English
            1
            edit-2
            1 year ago

            If it writes research papers, that research still has to come from somewhere. Even if the whole study was performed by AI itself, how would that deligitimise the research? Science isn’t art, it’s irrelevant who the performing agent is. (As long as it’s not stolen)

    • Communist
      link
      fedilink
      English
      4
      edit-2
      1 year ago

      It is very easy to get to those numbers if you don’t include the rate of false positives. That is all there is to this, really.