Here are two images of a house. There’s one obvious difference, but to this patient, P.S., they looked completely identical.
P.S. had suffered a stroke that damaged the right side of her brain, leaving her unaware of everything on her left side. But though she could discern no difference between the houses, when researchers asked her which she would prefer to live in, she chose the house that wasn’t burning— not once, but again and again.
P.S.’s brain was still processing information from her whole field of vision. She could see both images and tell the difference between them, she just didn’t know it. If someone threw a ball at her left side, she might duck. But she wouldn’t have any awareness of the ball, or any idea why she ducked.
P.S.’s condition, known as hemispatial neglect, reveals an important distinction between the brain’s processing of information and our experience of that processing. That experience is what we call consciousness. We are conscious of both the external world and our internal selves— we are aware of an image in much the same way we are aware of ourselves looking at an image, or our inner thoughts and emotions. But where does consciousness come from? Scientists, theologians, and philosophers have been trying to get to the bottom of this question for centuries— without reaching any consensus.
One recent theory is that consciousness is the brain’s imperfect picture of its own activity.
To understand this theory, it helps to have a clear idea of one important way the brain processes information from our senses. Based on sensory input, it builds models, which are continuously updating, simplified descriptions of objects and events in the world. Everything we know is based on these models. They never capture every detail of the things they describe, just enough for the brain to determine appropriate responses. For instance, one model built deep into the visual system codes white light as brightness without color. In reality, white light includes wavelengths that correspond to all the different colors we can see. Our perception of white light is wrong and oversimplified, but good enough for us to function. Likewise, the brain’s model of the physical body keeps track of the configuration of our limbs, but not of individual cells or even muscles, because that level of information isn’t needed to plan movement. If it didn’t have the model keeping track of the body’s size, shape, and how it is moving at any moment, we would quickly injure ourselves.
The brain also needs models of itself. For example, the brain has the ability to pay attention to specific objects and events. It also controls that focus, shifting it from one thing to another, internal and external, according to our needs. Without the ability to direct our focus, we wouldn’t be able to assess threats, finish a meal, or function at all. To control focus effectively, the brain has to construct a model of its own attention. With 86 billion neurons constantly interacting with each other, there’s no way the brain’s model of its own information processing can be perfectly self-descriptive. But like the model of the body, or our conception of white light, it doesn’t have to be. Our certainty that we have a metaphysical, subjective experience may come from one of the brain’s models, a cut-corner description of what it means to process information in a focused and deep manner.
Scientists have already begun trying to figure out how the brain creates that self model. MRI studies are a promising avenue for pinpointing the networks involved. These studies compare patterns of neural activation when someone is and isn’t conscious of a sensory stimulus, like an image. The results show that the areas needed for visual processing are activated whether or not the participant is aware of the image, but a whole additional network lights up only when they are conscious of seeing the image. Patients with hemispatial neglect, like P.S., typically have damage to one particular part of this network. More extensive damage to the network can sometimes lead to a vegetative state, with no sign of consciousness.
Evidence like this brings us closer to understanding how consciousness is built into the brain, but there’s still much more to learn. For instance, the way neurons in the networks related to consciousness compute specific pieces of information is outside the scope of our current technology. As we approach questions of consciousness with science, we’ll open new lines of inquiry into human identity.