• dan1101@lemmy.world
    link
    fedilink
    English
    arrow-up
    52
    ·
    edit-2
    2 years ago

    As expected, they can’t be trusted. And the more AI evolves, the less likely AI content will be detectable IMO.

    • jocanib@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      14
      arrow-down
      10
      ·
      2 years ago

      It will almost always be detectable if you just read what is written. Especially for academic work. It doesn’t know what a citation is, only what one looks like and where they appear. It can’t summarise a paper accurately. It’s easy to force laughably bad output by just asking the right sort of question.

      The simplest approach for setting homework is to give them the LLM output and get them to check it for errors and omissions. LLMs can’t critique their own work and students probably learn more from chasing down errors than filling a blank sheet of paper for the sake of it.

      • weew@lemmy.ca
        link
        fedilink
        English
        arrow-up
        31
        ·
        2 years ago

        given how much AI has advanced in the past year alone, saying it will “always” be easy to spot is extremely short sighted.

      • Zeth0s@lemmy.world
        link
        fedilink
        English
        arrow-up
        6
        ·
        2 years ago

        This is not entirely correct, in my experience. With the current version pf gtp-4 you might be right, but the initial versions were extremely good. Clearly you have to work with it, you cannot ask for the whole work

          • Zeth0s@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            ·
            edit-2
            2 years ago

            I meant initial versions of chatGTP 4. ChatGTP isn’t lying, simply because lying implies a malevolent intent. Gtp-4 has no intent, it just provides an output given an input, that can be either wrong or correct. A model able to provide more correct answers is a more accurate model. Computing accuracy for a LLM is not trivial, but gpt-4 is still a good model. User has to know how to use it, what to expect and how to evaluate the result. If they are unable to do so it’s completely their fault.

            Why are you so pissed of a good nlp model?

      • Asifall@lemmy.world
        link
        fedilink
        English
        arrow-up
        3
        ·
        2 years ago

        I think there’s a big difference between being able to identify an AI by talking to it and being able to identify something written by an AI, especially if a human has looked over it for obvious errors.