I’m not saying the record is a musician, I’m saying it can play music. And I’m not saying LLMs are sapient people, I’m saying they have a sense of self. An LLM is adept enough at adapting its copied idea of selfhood to its situation that it has a sense of self. It’s not as complex a sense a self as a human’s is, but it’s more complex than a magpie’s, and magpies pass the red dot mirror test of sentience. An LLM can adapt its copied ideas of self-awareness to the situation better than a magpie can.
im not arguing sapience, im examining your definition of sentience, which was self-awareness. my question was how we distinguish between mimicry of a sentient being and actually being sentient, with an analogy that a recording of a sentient being is a perfect mimicry but isn’t the same as having sentience.
similarly, how do we know that an llm is self aware and not merely a machine that returns clever combinations of recorded sentient beings? what is the equivalent of a red dot mirror test for an llm?
I’m not saying the record is a musician, I’m saying it can play music. And I’m not saying LLMs are sapient people, I’m saying they have a sense of self. An LLM is adept enough at adapting its copied idea of selfhood to its situation that it has a sense of self. It’s not as complex a sense a self as a human’s is, but it’s more complex than a magpie’s, and magpies pass the red dot mirror test of sentience. An LLM can adapt its copied ideas of self-awareness to the situation better than a magpie can.
im not arguing sapience, im examining your definition of sentience, which was self-awareness. my question was how we distinguish between mimicry of a sentient being and actually being sentient, with an analogy that a recording of a sentient being is a perfect mimicry but isn’t the same as having sentience.
similarly, how do we know that an llm is self aware and not merely a machine that returns clever combinations of recorded sentient beings? what is the equivalent of a red dot mirror test for an llm?