Sentience: Does AI Have Artificial Consciousness?

AI Artificial Intelligence

There is a recent article, Why Conscious AI Is a Bad, Bad Idea, where the author stated that, “While some researchers suggest that conscious AI is close at hand, others, including me, believe it remains far away and might not be possible at all. But even if unlikely, it is unwise to dismiss the possibility altogether. The prospect of artificial consciousness raises ethical, safety, and societal challenges significantly beyond those already posed by AI. Importantly, some of these challenges arise even when AI systems merely seem to be conscious, even if, under the hood, they are just algorithms whirring away in subjective oblivion.”

There is another recent article, Artificial Intelligence—Still a long way away, where the author stated that, “Researchers love to expound on the complexity of their learning models and how even they don’t understand what the models are actually doing and how it learns what it seems to do. And indeed, in some limited domains, such as learning how to play chess or go, machine learning can seem to learn how to do specific tasks much better than humans. But in the context of GENERAL artificial intelligence or AGIs (which is the kind that most pundits are afraid will eventually take over the world), machine learning is nowhere near the level that we can consider intelligent. And this breaks down to one simple fact: AGIs don’t actually know anything.”

The question that these articles, or others with similar drift, skip is: what does AI know? What do these large language models know? If they can respond to some queries or can carry out certain tasks accurately, those things, how can they be categorized?

Humans have external senses, but those are not ends. Often, it is see to know or see and know, applying to hear, touch, smell and taste. When something is sighted, its completion is what is known about it. If it is known, fine, if not, decision may vary.

During deep sleep, in a coma or under general anesthesia when it appears that knowing is AWOL, the activities that allow knowing to go on—in the mind—are present, even though their degree does not get to awareness or attention. The parallel to this are internal senses, where they function properly without awareness, until something changes or gets known, directly or referred, like pain.

A key difference in the moments after death, from the last moments of life, is that the ability for the individual to know has closed. Hearing, pain, and others are no longer known.

In general, feelings and emotions are not categorized with knowing, but when things are felt they can be said to be felt and known, like cold, heat, pain and thirst. The same applies to anger, delight, interest and love. The emotion is known. Subjective experiences are also known, for example the position of a limb, the location of an itch and so forth.

Knowing is central with a key source, the mind. If knowing were the brain, death may not be possible. But the mind, overseeing knowing, packs and life goes. The cells and molecules of the brain structure, organize, build or construct the mind. The mind, however, has a different structure, function and components from the brain. Serotonin is the brain, mood is the mind.

Organisms without a brain have a form—or mechanism—similar to the mind that helps them know. The organisms with a brain have a form of mind. The components of the mind, conceptually, are quantities and properties. Humans have more quantities and properties than others, or can know more than others—making humans dominant. Organisms can be said to be categorized by what they can know.

The list of what humans know includes perceptions, sensations, feelings, emotions, memory, and so forth. Memory includes creativity, reasoning, language, intelligence, skill and so on. All that humans know can be summed up to 1.

Plants and animals know, but their total is not 1. AI does not have feelings, emotions, sensations but it mirrors parts of the human memory. The total possible of what AI knows—or its memory, is higher than the share of memory for many plants and animals. This memory total is also possibly higher than the total knowing for some simple plants and animals.

LLMs may hallucinate, lack some logic, mix things up or not understand, but to the extent that they submit accuracies, they have a division of consciousness.

Consciousness can be defined as the ability for any system to know. Humans are the standard with a value of 1. The mind gives knowing or consciousness. The sentience of AI, like humans, is predicated on knowing, to the limited maximum—in the memory category.

David Stephen does research in theoretical neuroscience. He was a visiting scholar in medical entomology at the University of Illinois, Urbana-Champaign, UIUC. He did research in computer vision at Universitat Rovira i Virgili, URV, Tarragona.

Support Countercurrents

Countercurrents is answerable only to our readers. Support honest journalism because we have no PLANET B.
Become a Patron at Patreon

Join Our Newsletter

GET COUNTERCURRENTS DAILY NEWSLETTER STRAIGHT TO YOUR INBOX

Join our WhatsApp and Telegram Channels

Get CounterCurrents updates on our WhatsApp and Telegram Channels

Related Posts

Quest of immortality

Quest of immortality has been as timeless as it has remained fruitless. First recorded over 4000 years ago on the Babylonian Tablets, it narrates the tale of King Gilgamesh of Uruk. His…

Join Our Newsletter


Annual Subscription

Join Countercurrents Annual Fund Raising Campaign and help us

Latest News