• 29 Posts
  • 492 Comments
Joined 3 years ago
cake
Cake day: June 19th, 2023

help-circle









  • Andy@slrpnk.nettoMemes@lemmy.mlSupport Fediverse Thought Police
    link
    fedilink
    arrow-up
    3
    arrow-down
    4
    ·
    28 days ago

    Thanks for clarifying.

    At a glance, I don’t see a problem. Isn’t social media already a system for rating social credit?

    I think the problem with social credit scores is when they’re mandatory and can limit things like housing access. Filtering posts on opt-in social networks just sounds like a reasonable tool for moderating decentralized platforms.



  • This depends on your definition of self-awareness. I’m using what I think is a reasonable, mundane framework: self awareness is a spectrum of diverse capabilities that includes any system with some amount of internal observation.

    I think the definition that a lot of folks are using is a binary distinction between things which experience the ability to observe their own ego observing itself and those that don’t. Which I think is useful if your goal is to maintain a belief in human exceptionalism, but much less so if you’re trying to genuinely understand consciousness.

    A lizard has no ego. But it is aware of its comfort and will move from a cold spot to a warmer spot. That is low-level self awareness, and it’s not rare or mystical.









  • A hamster can’t generate a seahorse emoji either.

    I’m not stupid. I know how they work. I’m an animist, though. I realize everyone here thinks I’m a fool for believing a machine could have a spirit, but frankly I think everyone else is foolish for believing that a forest doesn’t.

    LLMs are obviously not people. But I think our current framework exceptionalizes humans in a way that allows us to ravage the planet and create torture camps for chickens.

    I would prefer that we approach this technology with more humility. Not to protect the “humanity” of a bunch of math, but to protect ours.

    Does that make sense?


  • Frankly I think our conception is way too limited.

    For instance, I would describe it as self-aware: it’s at least aware of its own state in the same way that your car is aware of it’s mileage and engine condition. They’re not sapient, but I do think they demonstrate self awareness in some narrow sense.

    I think rather than imagine these instances as “inanimate” we should place their level of comprehension along the same spectrum that includes a sea sponge, a nematode, a trout, a grasshopper, etc.

    I don’t know where the LLMs fall, but I find it hard to argue that they have less self awareness than a hamster. And that should freak us all out.