When considering the possibility of consciousness in artificial systems, there are at least two different approaches. One approach asks: How likely is it that current AI systems are conscious โ€“ and what needs to be added to existing systems to make it more likely that they are capable of consciousness? Another approach asks: What types of AI systems are unlikely to be conscious, and how can we rule out the possibility of certain types of systems becoming conscious?

In his research, Wanja Wiese pursues the second approach. โ€œMy aim is to contribute to two goals: Firstly, to reduce the risk of inadvertently creating artificial consciousness; this is a desirable outcome, as itโ€™s currently unclear under what conditions the creation of artificial consciousness is morally permissible. Secondly, this approach should help rule out deception by ostensibly conscious AI systems that only appear to be conscious,โ€ he explains. This is particularly important because there are already indications that many people who often interact with chatbots attribute consciousness to these systems. At the same time, the consensus among experts is that current AI systems are not conscious.


Embrace the charm of nostalgia with our Vintage-Inspired Contemplative Pooh-Bear Graphic T-Shirt. The “Winnie” in Winnie-the-Pooh was based on a Canadian Brown Bear, aka Ursus americanus, named Winnipeg.

The free energy principle

Wiese asks in his essay: How can we find out whether essential conditions for consciousness exist that are not fulfilled by conventional computers, for example? A common characteristic shared by all conscious animals is that they are alive. However, being alive is such a strict requirement that many donโ€™t consider it a plausible candidate for a necessary condition for consciousness. But perhaps some conditions that are necessary for being alive are also necessary for consciousness?

In his article, Wanja Wiese refers to British neuroscientist Karl Fristonโ€™s free energy principle. The principle indicates: The processes that ensure the continued existence of a self-organizing system such as a living organism can be described as a type of information processing. In humans, these include processes that regulate vital parameters such as body temperature, the oxygen content in the blood and blood sugar. The same type of information processing could also be realized in a computer. However, the computer would not regulate its temperature or blood sugar levels, but would merely simulate these processes.


Processingโ€ฆ
Success! You're on the list.

Most differences are not relevant to consciousness

The researcher suggests that the same could be true of consciousness. Assuming that consciousness contributes to the survival of a conscious organism, then, according to the free energy principle, the physiological processes that contribute to the maintenance of the organism must retain a trace that conscious experience leaves behind and that can be described as an information-processing process. This can be called the โ€œcomputational correlate of consciousnessโ€. This too can be realized in a computer. However, itโ€™s possible that additional conditions must be fulfilled in a computer in order for the computer to not only simulate but also replicate conscious experience.

In his article, Wanja Wiese therefore analyses differences between the way in which conscious creatures realize the computational correlate of consciousness and the way in which a computer would realize it in a simulation. He argues that most of these differences are not relevant to consciousness. For example, unlike an electronic computer, our brain is very energy efficient. But itโ€™s implausible that this is a requirement for consciousness.

Another difference, however, lies in the causal structure of computers and brains: In a conventional computer, data must always first be loaded from memory, then processed in the central processing unit, and finally stored in memory again. There is no such separation in the brain, which means that the causal connectivity of different areas of the brain takes on a different form. Wanja Wiese argues that this could be a difference between brains and conventional computers that is relevant to consciousness.

โ€œAs I see it, the perspective offered by the free energy principle is particularly interesting, because it allows us to describe characteristics of conscious living beings in such a way that they can be realized in artificial systems in principle, but arenโ€™t present in large classes of artificial systems (such as computer simulations),โ€ explains Wanja Wiese. โ€œThis means that the prerequisites for consciousness in artificial systems can be captured in a more detailed and precise way.โ€


Better cognition tied to higher relapse risk after depression remission
A study found that cognitive problems like memory loss may not predict …
New research suggests sexual arousal could blind people to rejection cues
Research shows sexual arousal can distort perceptions, leading individuals to misinterpret ambiguous …
Scientists stunned: Volcano cleans up after itself by removing methane from the air
The 2022 Hunga Tongaโ€“Hunga Haโ€™apai eruption unexpectedly cleared methane pollution, revealing a …
Early life on Earth relied on a surprisingly scarce metal
A study from UWโ€“Madison reveals that ancient life, 3.4 billion years ago, …

One response to “Can consciousness exist in a computer simulation?”

  1. It’s becoming clear that with all the brain and consciousness theories out there, the proof will be in the pudding. By this I mean, can any particular theory be used to create a human adult level conscious machine. My bet is on the late Gerald Edelman’s Extended Theory of Neuronal Group Selection. The lead group in robotics based on this theory is the Neurorobotics Lab at UC at Irvine. Dr. Edelman distinguished between primary consciousness, which came first in evolution, and that humans share with other conscious animals, and higher order consciousness, which came to only humans with the acquisition of language. A machine with only primary consciousness will probably have to come first.

    What I find special about the TNGS is the Darwin series of automata created at the Neurosciences Institute by Dr. Edelman and his colleagues in the 1990’s and 2000’s. These machines perform in the real world, not in a restricted simulated world, and display convincing physical behavior indicative of higher psychological functions necessary for consciousness, such as perceptual categorization, memory, and learning. They are based on realistic models of the parts of the biological brain that the theory claims subserve these functions. The extended TNGS allows for the emergence of consciousness based only on further evolutionary development of the brain areas responsible for these functions, in a parsimonious way. No other research I’ve encountered is anywhere near as convincing.

    I post because on almost every video and article about the brain and consciousness that I encounter, the attitude seems to be that we still know next to nothing about how the brain and consciousness work; that there’s lots of data but no unifying theory. I believe the extended TNGS is that theory. My motivation is to keep that theory in front of the public. And obviously, I consider it the route to a truly conscious machine, primary and higher-order.

    My advice to people who want to create a conscious machine is to seriously ground themselves in the extended TNGS and the Darwin automata first, and proceed from there, by applying to Jeff Krichmar’s lab at UC Irvine, possibly. Dr. Edelman’s roadmap to a conscious machine is at https://arxiv.org/abs/2105.10461

Leave a Reply

Trending

Discover more from Scientific Inquirer

Subscribe now to keep reading and get access to the full archive.

Continue reading