• Grail@multiverse.soulism.net
    link
    fedilink
    English
    arrow-up
    1
    ·
    18 hours ago

    I’m not saying the record is a musician, I’m saying it can play music. And I’m not saying LLMs are sapient people, I’m saying they have a sense of self. An LLM is adept enough at adapting its copied idea of selfhood to its situation that it has a sense of self. It’s not as complex a sense a self as a human’s is, but it’s more complex than a magpie’s, and magpies pass the red dot mirror test of sentience. An LLM can adapt its copied ideas of self-awareness to the situation better than a magpie can.

    • CrocodilloBombardino@piefed.social
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 hours ago

      im not arguing sapience, im examining your definition of sentience, which was self-awareness. my question was how we distinguish between mimicry of a sentient being and actually being sentient, with an analogy that a recording of a sentient being is a perfect mimicry but isn’t the same as having sentience.

      similarly, how do we know that an llm is self aware and not merely a machine that returns clever combinations of recorded sentient beings? what is the equivalent of a red dot mirror test for an llm?