I’m pulling the “twitter is a microblog” rule even though twitter is pretty mega now, hope that’s ok.

  • turdas@suppo.fi
    link
    fedilink
    English
    arrow-up
    1
    ·
    9 days ago

    No it can’t. If you’re actually saying that modern LLMs are no better at passing the Turing test than ELIZA, you are either trolling or an utterly delusional AI hater. Here, have a paper that proves you wrong: https://arxiv.org/pdf/2503.23674

    I am not saying the Turing test is a good benchmark of consciousness. On the contrary, like I said, LLMs have proven that it is not. But mere ten years ago even the most advanced chatbots had no hope of passing it, whereas now the most advanced ones are selected as the human over 70% of the time in a test that pits the LLM against a human head to head.

    • Nalivai@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 days ago

      No I’m saying the Turing test is a philosophical hypothetical from the time before computers, and doesn’t actually show anything, because it relies on the least accurate tool at our disposal: human pattern recognition machine, one that is oh so happy to be fooled by the ELIZAS of various sofistication. Chatbots were passing the Turing test since the invention of a chatbot. Yeah, modern chatbots are better at that, but it’s more of a damnation of our perception

      • turdas@suppo.fi
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 days ago

        OK, sounds like we broadly agree then.

        But as you can see in the paper I linked, ELIZA passes the Turing test in their experiment about 20% of the time (that is to say, it doesn’t pass; passing is 50% in this test) whereas the best LLMs pass about 70% of the time (that is to say, they are significantly more convincing at being human than real humans).

        • Nalivai@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          8 days ago

          That 20% figure is just a clear indication how shit people are at conducting such a test, and that was basically my original point. 2 in 10 times people were convinced by a particularly echoey room.

          • Fedegenerate@fedinsfw.app
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            8 days ago

            Turing test can be reliably passed by a bot that repeats last part of the previous sentence with a question mark at the end […]

            If an LLM is correct 2 in 10 times, would you call it “reliably correct”?

            • Nalivai@lemmy.world
              link
              fedilink
              English
              arrow-up
              1
              ·
              8 days ago

              If a person murders people only two days out of 10, they’re a murderer, in order to not be a murderer they need to never do that.
              Reliably correct is when you’re correct always. Demonstrably incorrect is when you’re incorrect even sometimes.

              • Fedegenerate@fedinsfw.app
                link
                fedilink
                English
                arrow-up
                1
                ·
                edit-2
                8 days ago

                Reliably correct is when you’re correct always.

                Agreed, except I add “almost”. “My car reliably starts” it starts “almost always”: more than 2 in 10 times. “You reliably turn up on time” doesn’t mean you’re late 8 in 10 times, it means you almost always turn up on time. To “almost always”, or “reliably” a thing: it means you fail 1 in 100, in a 1000, in 10,000 times. 10k is hyperbole, but the idea is clear right? Almost always/reliably != failing 8 out of 10 times.

                Your original point that these bots, that pass 2 in 10 times, reliably pass was wrong. Because: they dont “always pass”, they don’t “almost always” pass, they dont, even “pass in the majority of times”, they rarely pass.

                Let’s add our reliable = always substitution to the quote:

                Turing test can be [always] passed by a bot that repeats last part of the previous sentence with a question mark at the end […]

                You see how that’s wrong not just in fact, but in spirit too?

                If a person murders people only two days out of 10, they’re a murderer, in order to not be a murderer they need to never do that.

                Relevance? Who says “Fegenerate is reliably a murder?”

                Demonstrably incorrect is when you’re incorrect even sometimes.

                Relevance? You didn’t use the word "demonstrably passed’. I’d have no problems is you did?