OpenAI just admitted it can’t identify AI-generated text. That’s bad for the internet and it could be really bad for AI models.::In January, OpenAI launched a system for identifying AI-generated text. This month, the company scrapped it.

  • lily33@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    6
    ·
    1 year ago

    If you give me several paragraphs instead of a single sentence, do you still think it’s impossible to tell?

    • steakmeout@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      “If you zoom further out you can definitely tell it’s been shopped because you can see more pixels.”

    • steveman_ha@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      1 year ago

      What they’re getting towards (one thing, anyways) is that “indistinguishable to the model” and “the same” are two very different things.

      IIRC, one possibility is that LLMs which learn from one another will make such incremental changes to what’s considered “acceptable” or “normal” language structuring that, over time, more noticeable linguistic changes begin to emerge that go unnoticed by the models.

      As it continues, this phenomena creates a “positive feedback loop” in which the gap progressively widens – still undetected, because the quality of training data is going down – to the point where models basically “collapse” in their effectiveness.

      So even if their output is indistinguishable now, how the tech is used (I guess?) will determine whether or not a self-destructive LLM echo chamber is produced.