this post was submitted on 25 Sep 2023
175 points (84.9% liked)

Asklemmy

43965 readers
821 users here now

A loosely moderated place to ask open-ended questions

Search asklemmy ๐Ÿ”

If your post meets the following criteria, it's welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

~Icon~ ~by~ ~@Double_[email protected]~

founded 5 years ago
MODERATORS
 

Interesting article didnt know where it fit best so I wanted to share it here.

you are viewing a single comment's thread
view the rest of the comments
[โ€“] [email protected] 5 points 1 year ago (1 children)

No, consciousness is just what it feels like when a meat brain uses its meat to change its focus of attention; which gives rise to beliefs (some of them even true!) about a meat brain having a self.

It takes time, because brains are made of meat, and meat is slow.

It's leaky, because brains are made of meat, and meat oozes.

It generates the image of a "self" because brains are in meat bodies and actually do have physical continuity rather than being disconnected instants of computation; a term for "I, me, myself" is a rough model of the existence of brain features like memory, meat features like hormones, and even ape social-behavior features.

Attention/awareness is leaky and takes time; meat pumps rhythmically; and chemicals stick around.

And the meat brain can notice its own meaty doings. Just as it builds models of the outside world, it builds models of itself, with thoughts like "I am in the middle of doing an action" or "I am impatient" or "I feel sleepy" or "OW, LEG CRAMPS SUCK!" That is, its attention can range over not only the leg cramp itself, but its own reaction to having a leg cramp, including how the existence of leg cramps fits into its larger model of whether the world is a terrible place.

It usually comes up with a lot of correct beliefs out of this reflection, like "this is my leg, not your leg" and "I know English" and "Wow, I am distractable this morning, maybe it's the strong coffee". But it also comes up with dubious beliefs like "I am an eternal soul", "I am fully continuous in time", or "Oh God, what sin did I commit to deserve this leg cramp?"

("This is my leg, not yours" is important because there's nothing anyone can do to your leg that will make my leg cramp go away. The "self/other" distinction is important to consciousness because it has real-world implications; bodies really are physically disconnected from one another, which is why depersonalization can be an unhealthy thing for a consciousness to do too much.)

There's no reason to believe ChatGPT or the like are conscious, because they don't have the properties that consciousness is a model of. They're not fed information about their own well-being or place in the world. They don't observe their own processing. They do run largely as disconnected instants of computation. They don't live in a space where having a sense of "self/other" is effective.

(Not yet, anyway. There are folks out there trying to build AI systems that do have the feedback loops that might generate something like consciousness. This is probably a bad idea, and may even be an evil one.)

[โ€“] [email protected] 2 points 1 year ago

I like your reasoning very much, really well said!