Subtitles section Play video
[APPLAUSE]
- Two out of the three fundamental mysteries
about our place in the universe have already been resolved.
The first is literally about our place in the universe.
Many years ago Copernicus told us that we were not at its
centre, that we were just a tiny dot suspended in the abyss.
This is an image of the earth taken from the probe Voyager 1
as it was leaving the solar system from about six
billion kilometres away.
All of human history, all of the history of life on Earth,
has taken place on that pale blue dot.
The second mystery, Darwin then revealed
that we humans are just one branch, or one twig,
of a beautifully rich and delicate evolutionary tree.
And that much of the machinery of life
is shared even with the lowliest of our fellow creatures.
The third mystery is that of consciousness,
our inner universe.
Now earlier this year, for the third time in my life,
I ceased to exist.
As the propofol anaesthetic flowed from the cannula
in my wrist into my bloodstream and then into my brain,
there was a falling apart.
A blackness.
An absence.
And then, I was back.
Drowsy and disoriented, but definitely there.
And when you wake from a deep sleep,
you might be confused what time it is, especially
in flying somewhere, but you'll know that some time has passed.
There seems to be some basic continuity
between your consciousness then, and your consciousness now.
But coming around from a general anaesthetic,
it could have been five minutes.
It could have been five hours.
It could have been five days, or five years.
I was simply not there.
A premonition of the oblivion of death.
And general anaesthesia doesn't just work on your brain.
It doesn't just work on your mind.
It works on your consciousness.
By altering the delicate electrochemical circuitry
inside your head, the basic ground state
of what it is to be is temporarily abolished.
And in this process lies one of the greatest remaining
mysteries in science and philosophy.
How does consciousness happen?
Why is life in the first person?
It is going away, and coming back.
The modern incarnation of this problem
is usually traced to Descartes, who in the 17th century
distinguished between matter stuff, res extensa, the stuff
that these desks are made of, that clothes are made of.
But also the brains and bodies and made of, material stuff.
And res cogitans, the stuff of thought, of feelings.
The stuff of consciousness.
And in making this distinction, he gave rise
to the now infamous mind/body problem,
and life has never been simple ever since.
But Descartes actually generated even more mischief
with his doctrine of the beast machine,
which I'm going to mention now, because it anticipates where
I'm going to end up as the bell rings when I finish in an hour.
Before Descartes, people commonly believed in something
called the great chain of being, with rocks and plants
at one end, and other non-human animals, a bit higher
up than humans, and then angels and gods at the very top.
And this great scale of being was also
a scale of moral virtue, so that humans had more moral virtues
than animals and plants, and then rocks and so on.
Now Descartes, in making this division between mind
and matter, argued that only humans had minds, and therefore
moral status, while other animals didn't have minds.
They were merely physiological machines, or beast machines,
morally equivalent to plants, and to rocks.
And in this view, the physiological mechanisms
that give rise to the property of being alive
were not relevant to the presence
of mind or consciousness.
Now I'm going to propose, at the end of this talk, the opposite.
That our conscious sense of self arises because of, and not
in spite of, the fact that we, too, are beast machines.
So to get there, let's return to the apparent mystery
of consciousness.
Now as recently as 1989, which is quite a while ago, but not
that long ago, Stuart Sutherland,
who was founding professor of experimental psychology
at my university of Sussex, had this to say.
"Consciousness is a fascinating but elusive phenomenon.
It is impossible to specify what it is, what it does,
or why it evolved.
Nothing worth reading has been written on it."
[LAUGHTER]
It's quite a pessimistic point of view.
And that may have been true then.
I don't think it was true then, but in any case
things have changed a lot since.
And more or less, about the time that Sutherland
made these remarks, we can see the birth, or the rebirth,
of the study of consciousness within the neurosciences.
And a good landmark is this paper
by Francis Crick and Christof Koch, published in 1990.
And they start their paper by saying that it is remarkable
that most of the work in cognitive sciences,
and the neurosciences, makes no reference to consciousness
or awareness at all.
And then they go on to propose their own theory
of what the neural correlates of consciousness are.
What it is in the brain that is responsible for being
conscious.
And since then, over the last 25 years,
there's been first a trickle, and now a deluge of research
on the brain basis of conscious experience.
Some of this work is being carried out
in my laboratory, the Sackler Centre, the consciousness
science, which was founded six years ago with Hugo
Critchley, my co-director.
And there are now even specialised academic journals,
The Neuroscience of Consciousness,
which I started last year with Oxford University Press.
And this is a real change of the tide.
When I started out more than 20 years ago,
it was thought to be a very-- it was thought
to be career suicide to want to study consciousness,
scientifically.
And it may still be, we don't know.
Let's see.
So while the brain basis of consciousness
is still a mystery, it is, in some sense,
an accessible mystery.
And the author, Mark Haddon, put this very nicely, I think.
He said the raw material of consciousness
is not on the other side of the universe.
It didn't happen 14 billion years ago.
And it's not squirrelled away deep inside an atom.
The raw material of consciousness
is right here, inside your head, and you can
hold the brain in your hands.
But the brain won't deliver its secrets very easily.
What's extraordinary about the brain
is not so much the number of neurons,
though there are about 90 billion.
It's not even the number of connections,
though there are so many that if you counted one every second,
it would take you about three million years
to finish counting.
What's truly extraordinary are the patterns
of connectivity, which to a large extent,
are still not known, but within which are inscribed everything
that makes you, you.
The challenge is then this, at least the the way I see it.
How can the structure and dynamics
of the brain, in connection with the body and the environment,
account for the subjective phenomenological properties
of consciousness?
And considering things this way, we
come up against what the philosopher David
Chalmers has often called the hard problem of consciousness.
And the idea is this.
There is an easy problem.
The easy problem is to understand how the combined
operations of the brain and the body give rise to perception,
to cognition, to thinking, to learning, to behaviour.
How the brain works, in other words.
The hard problem is to understand
why and how any of this should have anything
to do with consciousness at all.
Why aren't we just robots, or philosophical zombies,
without any in a universe?
Now there's a tempting intuition here,
which is that, even if we solve the hard problem, even if we
solve the easy problem, the hard problem would still remain
as mysterious as it seems now.
But this just seems wrong-headed to me.
It may not be necessary to explain why consciousness
exists at all, in order to make progress in understanding
its material basis.
And this for me, is the real problem of consciousness;
how to account for its various properties in terms
of biological mechanisms without pretending that it doesn't
exist at all, as you do if you solve the easy problem,
and without trying to account for why it's
parts of the universe in the first place, which
is the hard problem.
And in the history of science, we've
been somewhere similar before.
It's hard to say if it's exactly the same situation.
But in our understanding of life,
eminent biochemists of the time found it entirely mysterious
how biological mechanisms could give rise
to the property of being alive.
And there were proposed of things,
like elan vital and essence vital,
and all sorts of other stuff.
And although we don't yet understand everything
about life, this initial sense of mystery about life
has, to a large extent, dissolved
as biologists have just got on with the business
of understanding the properties of living systems in terms
of mechanisms.
An important part of this story was
the realisation that life is not just one thing,
but rather a constellation of partially dependent, partially
separable, processes, like metabolism, homeostasis,
and reproduction.
In the same way, to make progress
on the real problem of consciousness,
it can be useful to distinguish different aspects or dimensions
of what it is to be conscious.
The space of possible minds, if you like.
And one simple classification is into conscious level,
which is the property of being conscious at all.
For example, the difference between being
in a dreamless sleep, or under general anaesthesia,
and being awake and conscious as you are now.
And the conscious content, when you are conscious,
you're conscious of something.
The myriad of sights, sounds, smells, emotions, feelings,
and beliefs that populate your inner universe at any one time.
And one thing you are conscious of when you are conscious,
is the experience, the specific experience, of being you,
and this is conscious self.
And it's the third dimension of consciousness.
Now I don't claim these distinctions mark
completely independent aspects of what it is to be conscious,
but they're a pragmatically useful way of breaking down
the problem a bit.
So let's start with conscious level.
What are the fundamental brain mechanisms
that underlie our ability to be conscious at all?
And we can think of this, at least a first approximation,
as a scale from being completely unconscious,
as if you were in a coma, or under general anaesthesia,
to being awake, alert, and fully conscious as you are now.
And there's various states in between being drowsy,
being mildly sedated and so on.
What's important is that, while being conscious and being awake
often go together, this is not always the case.
For instance, when you are dreaming you are asleep,
but you are having conscious experiences.
The conscious experience of your dreams.
And on the other side of this diagram,
there are pathological states, like the vegetative state,
where physiologically you will go through sleep/wake cycles,
but there is nobody at home.
There is no consciousness happening.
So what are the specific mechanisms
that underlie being conscious and not simply being
physiologically awake?
Well there are a number of possibilities.
Is it the number of neurons?
Well actually, probably not.
There are more neurons in your cerebellum,
this bit at the back of your brain,
than in the rest of your brain put together.
In fact are about four times more neurons in your cerebellum
than in the rest of your cortex.
But if you have damage to your cerebellum, yeah,
you'll have some problems with coordination
and other things, some cognitive problems,
but you won't lose consciousness.
It's not just the number of neurons.
Doesn't seem to be any particular region.
In fact, there are regions that, if you suffer damage,
you will permanently lose consciousness;
in thalamina, the nuclei, and the thalamus
deep inside the brain.
But these seems to be more like on/off switches
than actual generators of conscious experience.
It's not even neural activity, at least not
simple kinds of neural activity.
Your brain is still highly active
doing unconscious states, during the sleep.
And even if your brain is highly synchronised,
one of the first theories of consciousness
was it depended on neurons firing in synchrony
with each other.
If your brain is too synchronised,
you will lose consciousness, and this happens
in states of absence epilepsy.
What seems to be the case is that, being conscious it all,
depends on how different brain regions talk
to each other in specific ways.
And this was some groundbreaking work by Marcello Massimini
in Milan about 10 years ago.
And what he did here, was he stimulated
the cortex of the brain with a brief pulse
of electromagnetic energy, using a technique called
transcranial magnetic stimulation or TMS.
And then he used EEG electroencephalography
to listen to the brain's echos.
A little bit like banging on the brain
and listening to its electrical response.
And what he noticed when you do this,
and you can see on the left is asleep,
and on the right is awake.
And this is very much slowed down.
When you stimulate the brain in a sleep condition,
there is still a response.
There's still an echo, but the echo stays very localised
to the point of stimulation.
It doesn't travel around very much,
and it doesn't last very long.
But when you stimulate a conscious brain,
there's a spatial temporally complex response.
This echo bounces around all over the cortex
in very interesting ways.
What's more, the complexity of this echo can be quantified.
You can apply some simple algorithms
to describe how complex, how rich,
this pattern of interactivity is.
This is also from the Milan group.
And what they've done here is, they basically look at the echo
as it moves around the brain.
And they see the extent to which you could describe it,
the minimum description length.
How much can you compress the image of that echo?
Much the same way that algorithms make
compressed files from digital images in your phone.
And they came up with an index called
the perturbational complexity index.
And what you find is, you now have
a number that you can attach to how conscious you are.
This is, I think, really intriguing,
because it's a first step towards having
an actual measurement of conscious level.
This graph on the bottom shows this measure
applied to a variety of conscious states,
ranging from pathological conscious states,
like the vegetative state, where there
is no consciousness at all, all the way through locked
in syndrome, and then healthy waking.
And you can immediately see that techniques
like this might already have clinical value in diagnosing
potential for consciousness patients
might have after severe brain injury.
Now at Sussex, we are continuing work along these lines.
We actually look, instead of bashing
on the brain with a sharp pulse of energy,
we want to see whether we can get something similar just
by recording the spontaneous activity of the brain.
So we look at spontaneous dynamics
from, in this case, waking states and anaesthesia.
This is work with my PhD students Michael Schartner
and Adam Barrett.
We measure its complexity, and indeed, we
find that we can distinguish different levels
of consciousness just by the spontaneous activity
of the brain.
In a way this isn't that surprising,
because we know various things change.
The balance of different frequencies
of your brain activity changes when you lose consciousness.
But this doesn't have to do with that.
This is independent of that.
There's something specific that is being detected
by these changes in complexity.
More recently we've applied the same measures to sleep,
in this case taking advantage with colleagues
in Milan of recordings taken from directly
within the human cerebral cortex.
These are implanted electrodes.
And we see much the same story.
If you compare where the two areas are,
you compare the complexity of wakeful rest,
and early non-rem sleep, where you are not dreaming very much.
You see that complexity fools a great deal.
What's interesting here is, if you compare wakeful rest to REM
sleep, where people will often report dreams if you wake them
up, the level of complexity is very much
as it is during the wakeful state.
There's something else going on here,
which is that the complexity in the frontal part of the brain
seems to be higher than in other parts of the brain.
And that's something we still don't understand fully, yet.
I just wanted to give you something
hot off the press, so to speak, which is where you've also
been applying these measures now to data taken
from people under the influence of psychedelic drugs;
psilocybin, ketamine, and LSD.
And what we find, at least in our hands to start with here,
is that the level of complexity actually
increases as compared to the baseline state, which is not
something we've seen before in any other application
of these measures.
So what's important about this way of looking at things
is that, it's grounded in a theory that
tries to explain why certain sorts of brain dynamics
go along with being conscious.
And put very simply, the idea is this--
and it goes back to Guilio Tononi and Gerald Edelman,
people that I went to work with in America about 18 years ago--
the idea is very simple.
Consciousness is extremely informative.
Every conscious experience you have, or have had,
or will have, is different from every other conscious
experience you have had, are having, or will have.
Even the experience of pure darkness
rules out a vast repertoire of alternative possible
experiences that you could have, or might have in the future.
There's a huge amount of information for the organism
in any conscious experience.
At the same time every experience that you have
is highly integrated.
Every conscious scene is experienced, as all of a piece,
is bound together.
We don't experience colours and shapes separately in any way.
It's conscious experiences at the level of phenomenology
combine these properties.
They are the one hand highly informative, composed
of many different parts.
On the other, bound together in a unified whole.
And this motivates us to search for mathematical measures which
have the same property, which are neither
lacking in information.
On the left, you see a system which
is all connected together, so it can't enter
very many different states.
On the right is a system which is completely dissociated,
so they can enter states, but it's not a single system.
We want measures that track this middle ground of systems,
that combine both integration and differentiation.
And a number of these measures now exist.
There are some equations here, which
we can talk about later if you like,
that try to target this middle ground.
And time will tell whether, by applying these more
phenomenologically grounded measures,
we come up with even more precise practical measures
of consciousness.
Now why is this business of measurement important?
And I want to make a general point here,
which is that, if you're trying to naturalise
a phenomenon which seems mysterious,
the ability to measure it is usually one of the most
important steps you can take.
And we've seen numerous examples of this.
The history of our understanding of heat and temperature
is one very good example Here's an early thermometer
from the 18th century, which used the expansion of air.
But of course there are many problems
in generating a reliable thermometer
and a scale of temperature, if you don't already
have fixed points.
And if you don't know what heat is you
get trapped in a kind of vicious circle
that took a long time to break out of.
But people did break out of this,
and the development of the thermometers
catalysed our understanding of heat
from being something that flowed in and out of objects, to being
something that was identical to a physical property.
The mean molecular kinetic energy of molecules
in a substance.
And having that concept of heat now
allows us to talk about temperature
far beyond the realms of human experience.
We can talk about the temperature
on the surface of the sun, in a sense,
the way we can talk about the temperature of interstellar
space, close to absolute zero.
None of these things make any sense
and in our phenomenological experience of hot and cold.
So this brings me to my first take-home message.
Measurement is important, and consciousness, conscious level,
depends on a complex balance of differentiation and integration
in brain dynamics, reflecting the fact
that conscious experiences themselves
are both highly informative and always integrated.
Now when we are conscious, we are conscious of something.
So what are the brain mechanisms that determine
the content of consciousness?
And the hero for this part of the story
is the German physicist and physiologist Hermann Von
Helmholtz.
And he proposed the idea that the brain is
a kind of prediction machine.
That what we see, hear, and feel are
nothing other than the brain's best guess about the causes
of sensory inputs.
And the basic idea is, again, quite simple.
The brain is locked inside its bony skull home,
and has very indirect access to the external world.
All it receives are ambiguous and noisy sensory signals,
which are highly and directly related
to this external world of objects, and so on,
if there is an external world of objects out there at all.
They know about that.
Perception in this view is, by necessity,
a process of inference in which the brain interprets
these ambiguous and noisy sensory signals with respect
to some prior expectations or beliefs
about the way the world is.
And this forms the brain's best guess
of the causes of the sensory signals that
are impacting our sensory surfaces all the time.
What we see is the brain's best guess of what's out there.
I want to give you a couple of examples
that illustrate this process.
It's quite easy to do, in a way.
This first example is a well-known visual illusion
called Edelstein's Checkerboard.
Now here, you'll see two patches.
You'll see patches A and B. And I hope to you they look to be
different shades of grey.
Do they?
They look to be different shades of grey.
Of course they are exactly the same shade of grey.
I can illustrate that by putting an alternative image,
and joining up those two patches.
You'll see that's it's the same shade of grey.
You may not believe me, so what I'll do
is, I'll shift it along, and you'll see even more clearly.
There are no sharp edges.
It's the same shade of grey.
What's going on here, of course, is
that the brain is unconsciously applying its prior knowledge
that a cast shadow dims the appearance of surfaces
that it casts onto.
So we therefore see the patch B as being
lighter than it really is, in order
to account for that effect.
And this is of course an illustration
of the success of the visual system, not its failure.
The visual system is a very bad physical light metre,
but that's not what it's supposed to do.
It's supposed to, or one thing it's supposed to do,
is to interpret the causes of sensory signals
in terms of meaningful objects in the world.
It's also an example of what we sometimes
call cognitive impenetrability.
Even if you know the patches are the same shade of grey,
when I take that bar away, they again look different.
Can't do much about that.
The second example just shows you
how quickly the brain can take in new prior information
to change the nature of conscious perception.
This is a so-called Mooney image.
And if you haven't seen it before, hopefully what
you will see here is just a passing
of black and white splotches.
Does everybody kind of get that?
Black and white splotches?
Some of you might have seen this before.
And now what I'm going to do is fill it in,
and you see something very different.
What you'll see is a very meaningful scene here involving
at least two objects, a woman, a hat and a horse.
Now if you stare at this for a while,
I won't leave it up for too long--
but if you just look at it for a little bit,
and then I take that image away again,
you should still be able to see the objects within that image.
Now for me this is quite remarkable,
because the sensory information hasn't changed at all.
All that's changed are your brain's prior expectations
about what that sensory data signifies.
And this changes what you consciously see.
Now this also works in the auditory domain.
Here are two spectrograms.
This is something called sine wave speech,
and what you see here are two time frequency representations
of speech sounds.
The one on the top has all the sharp acoustical features that
provide normal speech removed.
A little bit like thresholding an image.
And the bottom is something else.
So I'm going to play the top first,
and let's see what it sounds like.
[STRANGE BEEPS AND NOISES]
And now I'll play you something else.
(BOTTOM SOUND - A MAN'S VOICE): Jazz and swing fans
like fast music.
- So I hope you all understood that piece of sage advice.
And now if I play the original sound again--
[BEEPS AND WHISTLES THAT SOUND LIKE THE SENTENCE]
- Yeah?
This is exactly the same.
Again, all this change is what we
expect that sound to signify.
[SAME SOUND PLAYED AGAIN]
- One more time, just for luck.
It's not just a bunch of noisy whistles, it's speech.
Now this typical framework for thinking
about these kinds of effects is Bayesian Inference.
And this is a form of probabilistic reasoning, which
is applicable in all sorts of domains,
not just in neuroscience, in medical diagnosis,
and all sorts of things, like finding lost submarines.
But in neuroscience, we talk about the Bayesian brain.
And it's a way of formalising Helmholtz's idea
that perception is a form of best guessing.
And the idea is that sensory signals and prior beliefs
can be represented as probability distribution.
So for instance, this yellow curve
is the probability of something being the case,
maybe that you've got a brief glimpse of an object moving
to the right.
The sensory data may say something different.
It may have a probability that peaks
at a different angle of movement.
Maybe it's drifting in a different direction.
And the optimal combination of the prior,
and the likelihood, the yellow curve and the red curve,
is this green curve, which we will call the posterior
distribution.
And that represents the best optimal combination
of these two sorts of evidence.
And the idea is, well that's what we perceive.
Thinking about perception in this way
does something rather strange to the way,
classically in neuroscience, people
have thought about perception.
The classical view is that the brain
processes sensory information in a bottom-up,
or feed-forward direction.
This is a picture of the visual system of the monkey,
and the idea is that information comes in through the retina,
then goes through the thalamus.
It then goes to the back of the brain.
And as the sensory signals percolate deeper and deeper
and deeper into the brain, they encode
or represent progressively more sophisticated features
of objects.
So you start out at early levels the visual cortex
with response to luminance, and edges,
and then higher up to objects, including other monkeys.
What's important here is the perceptual heavy lifting
is done by information flowing in this bottom-up or
feed-forward direction.
Now the Bayesian brain idea says something very different.
It says that what's really important are the top-down
or inside-out connections that flow from the centre
of the brain back out.
And we've known for a long time there's
a large number, a very large number of these connections,
and some descriptions more than flow the other way around.
But the function has been rather mysterious.
Thinking about the Bayesian brain
gives us a nice way to interpret this.
Which is that it's exactly these top-down or inside-out
connections that convey predictions
from high levels of the brain to lower levels, to lower levels,
back out to the sensory surfaces.
So these blue arrows convey the brain's predictions
about the causes of sensory signals.
And then what flows in the feed-forward or bottom-up
direction, from the outside-in, that's just the prediction
area, the difference between what the brain expects
and what it gets at each level of description.
So this is often called predictive coding,
or predictive processing, in some formal frameworks.
And the idea is that minimization of prediction
error occurs across all levels of this hierarchy
at the same time, and what we then
perceive is the consequence of this joint minimization
of prediction error.
So you can think of perception as a sort
of controlled hallucination, in which
our perceptual predictions are being reined in at all points
by sensory information from the world and the body.
Now there's quite a lot of experiments
that show that something like this
is probably going on in the brain.
These are a couple of examples.
And since they're--
I was looking for the best example
so they don't come from my lab at all.
[LAUGHTER]
This is from Lars Muckli in Glasgow.
He's shown using advanced brain reading techniques, which
I won't describe, that you can decode the context of what
a person is seeing from parts of the visual cortex that
isn't even receiving any input.
And what's more, you can decode better
when you decode from the top part of the cortex, which
is supposed to receive predictions from higher levels.
So that suggests there are predictions being fed back.
And another study by Andre Bastos and Pascal Fries
in Germany, they used a method called Granger causality, which
is sensitive to information flow in systems.
And they find that top-down signals and bottom-up signals
are conveyed in different frequency bands
in the cortex, which is what you'd
predict from predictive coding.
One last experiment which I find particularly interesting
is an experiment from a Japanese group of Masanori Murayama.
And they used to optigenetics, which
is a way of using lights to selectively turn
on or off neural circuits.
And in this experiment they showed
that by just deactivating top superficial levels
of somatosensory cortex in a mouse
brain, the part of the mouse brain
that's sensitive to touch, they could
affect how well that mouse was able to do
tactile discriminations.
Those top-down connections were coming from a motor cortex.
So there's a lot of evidence that top-down connections
in the brain are important for perception,
is the basic message there.
But what's rather strange, and what I'm going to tell
you next is that all this stuff is all very good,
but predictive processing is not a theory of consciousness.
Nothing I've said has anything to do with consciousness,
at all.
It has to do-- it's a very general theory of how
brains do what they do.
How they do perception, how they do
cognition, how they do action.
So somewhat counter-intuitively, I
think this is exactly why it's a great theory of consciousness.
And the reason I think this is because it
allows us to ask all sorts of questions
about the real problem.
About what it is, what happens in brains
that underlies what you happen to be conscious of right now,
without getting sucked into the metaphysical pluckhole of why
you are conscious in the first place.
In other words, it provides a powerful approach
to looking for neural correlates of consciousness,
those things in the brain that go along with being conscious.
Because we can now take advantage
of a very general theory of how brains do what
they do, rather than just looking
at this region or that region.
So what does predictive processing,
or the Bayesian brain say about consciousness, specifically?
Well many years ago, some influential experiments
revealed a very strong connection
between top-down signalling and conscious contents.
In this example by Alvaro Pascual-Leone and Vincent
Walsh, what they did was they had
people look at visual motion, examples of visual motion.
And they used TMS, this interventional technique
where you can zap the brain very briefly.
I mentioned it before.
But they used it here, specifically
to interrupt the top-M down signalling
that was evoked by this perception of visual motion.
And the result was that, if you interrupted specifically
the top-down feedback, you would abolish
the conscious perception of visual motion,
even if you left the bottom-up signalling intact.
So that was an early key.
Now, more recently, in our lab and in many other labs
all over the place, we've been asking some other questions
about the relationship between what you expect
and what you consciously experience.
One of the most basic questions you can ask
is, do we consciously see what we expect to see?
Or do we see what violates our expectations of what we expect?
And a recent study from our group, led by Yair Pinto,
used a method called continuous flash suppression
to address this question.
It's illustrated here.
You see different images in the different eyes.
In one eye you see this rapidly changing Mondrian pattern
of squares.
And in the other eye, you either see a face or a house.
And they change contrast like this.
So initially, the person would just see this random pattern,
and then they'll see either a house or a face.
And simply, you just ask them to expect
to see-- you just tell them a face is more likely
or a house is more likely.
And what we find over a number of studies
is that we see faces more quickly when that's
what we're expecting to see.
It may seem obvious, but it could be the other way around.
At least in these studies, we see what we expect to see,
not what violates our expectations.
That's the data.
And the same goes for houses.
These kinds of studies support the idea
that it's the top-down predictions that
are really important for determining
what we're conscious of.
There's another experiment which I will just mention.
We did pretty much the same thing.
This is called motion induced blindness.
If you are in a lab rather than in a lecture theatre,
and you stare at this central point here,
then this red dot might disappear from time to time.
And what we did was after it disappeared,
we changed its colour and we led people to expect the colour
change to be one thing or another.
And again, it reappeared more quickly if it changed colour
in the way you were expecting.
Again, I am confirming that once your expectations were
validated, then that accelerated your conscious awareness
of something in the world.
Now that's just behavioural evidence.
That's just asking people what they see and when they see it.
We've also been interested in the brain mechanisms that
underlie and shape how our expectations change,
what we consciously see.
And we've been particularly interested in something
called the alpha rhythm.
Now the alpha rhythm is an oscillation
of about 10 hertz or 10 cycles per second.
That's especially prominent in the visual cortex,
across the back of the brain.
In one study, led by in this case a PhD student,
Maxine Sherman, with Ryota Kanai, in Sussex.
What we did here, we manipulated people's expectations
of what they we're likely to see.
And it's a very boring experiment, this was.
The only thing they could see was what we call Gabor patches.
They're just very dim patches of lines that
are blurry around the edges.
But the visual system loves these kinds of things.
They activate early visual cortex very, very well.
And people were expecting either that a patch
was there, or that it wasn't there, in different conditions.
And while doing this we measured brain activity.
And to cut a long story short, what we found
was that there were certain phases, certain parts
of the cycle, this 10 Hertz cycle,
at which your expectation had a greater effect on what
you said that you saw.
So there was part of this cycle, as the alpha rhythm,
there was part of it where your expectations dominated
your perception.
And there was another part of it which
was the opposite, where your sensory signals were
more important in determining what you saw at that point.
So this suggests that this oscillation
in the back of the brain is orchestrating
this exchange of predictions and prediction errors.
And that's the sort of cycle that
might be the neural mechanism for conscious vision.
And other theories about what the alpha rhythm is doing,
there are many.
One is that it's doing nothing, it's just the brain
idling away, and I think this is at least a more interesting way
to think about it.
Another experiment we've done with another PhD student
Asa-Chang, we showed people these very fast
changing luminance sequences.
And it turns out that your brains
will learn to predict the specific changes
in these sequences that change you very quickly.
And the signature of this learning,
again, happens to be in the alpha rhythm,
and suggests that this oscillation has something
important to do with how the brain learns and updates
predictions about sensory signals.
But we do not go around the world looking at Gabor patches
or rapidly changing things like this.
We go around the world looking at people and objects.
And that's what our visual world is composed of.
So can any of these ideas say anything
about our everyday visual experience?
And I think that's a very important challenge
in neuroscience to cross.
Get out of the lab and think about real world experiences.
So we've been using virtual reality over the last few years
to try to get at some of these ideas.
This is an Oculus Rift, which is now available to buy, I think.
And we've been using these to address
some of these real world aspects of visual perception.
And one of these real world aspects
is called perceptual presence.
And this is the observation that,
in normal perception, objects really seem to be there,
as opposed to being images of objects.
And this is, of course, what Magritte
plays with in his famous painting,
The Treachery of Images.
For instance, this is an object.
I think it's there, and in some sense
I can perceive the back of it, even though I
can't see the back of it, even though the back of it's
not giving me any sensory data, I perceive it
as an object with a back.
How does one explain that?
Well one idea you can come up with within this Bayesian brain
framework is that, the brain is not only
predicting the possible causes of the sensory signals getting
right here, right now.
But it's also predicting how sensory signals
would change were I to make particular actions.
Were I to pick this object up and move it around,
or just move my eyes from one place to another.
There's a long paper.
I wrote about that which I--
please don't read it.
[LAUGHTER]
- But that's the basic idea.
And how do you test an idea like that?
So we've been using some innovative virtual reality
methods, or augmented reality methods,
with my post-doc, Keisuke Suzuki.
And what we do is, we have virtual objects,
and these virtual objects, they either behave
as a normal object would.
They're all weird, unfamiliar objects.
But they can either behave as a normal object would behave,
so you can learn to predict what would happen.
This one is weird.
It always shows you the same face,
a little bit like having the moon on a plate
in front of you.
And then there are other conditions
where objects respond to your movements,
but they do so in unreliable and strange ways.
So the question is, what does the brain
learn about these objects, and how do we experience them?
Do we experience them as objects in different ways
when they behave differently?
And we're still doing those experiments.
Another way we can use VR, is to investigate
what happens in visual hallucinations of the kind
experienced in psychosis, and in certain other more
pharmacologically induced conditions.
What we're doing here is, we're coupling
immersive virtual reality, imagine
you've got a headset strapped to your head
again, with clever image processing that models
the effects of overactive priors on visual perception
to generate a highly immersive experience.
This is Sussex campus, actually, but now it
seems quite different than it did at lunchtime today.
I'll tell you that.
What we've done, we've recorded this panoramic video
which we can feedback through VR headset.
And we've processed this video through one
of these Google deep dream algorithms you might have seen,
that can generate sort of bowls of pasta
that looks like animals.
And this might seem like a lot of fun.
It is fun, but there is a serious purpose here,
because it allows us to model the effects, model very
unusual forms of perception and how
they might play out in different ways in the visual hierarchy.
And understanding how visual hallucinations might happen,
and how the wider effects they have on the mind, I think,
is a very important part of studying visual perception.
So that brings us to the second take-home message,
which is that what we consciously see
is the brain's best guess of the causes of its sensory input.
Normal perception is a fantasy that is constrained by reality.
Now before I move on to the last section,
I want to pay tribute to an unlikely character in a talk
about neuroscience, which is Ernst Gombrich.
Ernst Gombrich was one of the foremost historians
of art of the 20th century.
And it turns out that Gombrich's approach to understanding art
had a lot in common with ideas and the Bayesian brain.
And more specifically, with the idea
that perception is largely an act
of imagination, or construction, on the part of the perceiver.
And this is most apparent in his concept of the beholder's
share, which emphasises that the viewer brings
an awful lot to the table in the act of experiencing an artwork.
So he had this to say in his 1960 book, Art and Illusion,
"the artist gives the beholder 'more to do', he draws them
into the magic circle of creation and allows him
to experience something of the thrill of making which had once
been the privilege of the artist."
I think for me this is very powerful
when looking at, especially, things like Impressionist art.
And here, one way to think about this
is, that the artist has reverse engineered
the whole perceptual process, so that what's
there are not the objects, the end points of perception,
but rather the raw materials; the patterns of light that
engage our perceptual machinery in doing its work.
And for me this might be why paintings like this
are particularly powerful.
Now the final dimension of consciousness
I want to talk about is conscious self.
The fundamental experience of being someone.
Being someone like you.
There are many aspects to our experience
of being a conscious self.
There is the bodily self, the experience
of being and identifying with a particular body.
A bit of the world goes around with you in the world
all the time.
There's the perspectival self, the experience
of seeing the world, or experiencing the world,
from a particular first person perspective,
usually somewhere in the body, but not always.
There is the volitional self, the experience
of intending to do things, and of making things
happen in the world of agency.
And these ideas are, of course, often associated
with concepts of will.
Then there's the narrative self.
This is where-- only until now, we
don't have to worry about the concept of I,
but when we get to the narrative self, there is now and I.
There is a continuity of self experience from hour to hour,
from day to day, from month to month,
and from year to year, that you associate a name with,
and a particular set of autobiographical memories.
And finally, there's a social self.
The way I experience being me is partly
dependent on the way I perceive you as perceiving me.
I'm just going to talk, in the minutes remain,
about the bodily self.
This is something we're working on quite a lot in Sussex.
The experience of identifying with, and owning,
a particular body.
And the basic idea I want to convey, is again, very simple.
It's just that we should think of our experience of body
ownership in the same way that we
think about our experience of other things, as well.
That is, it's the brain's best guess of the causes
of body-related signals.
And the brain is always making this inference.
It's making its inference about what in the world
is part of the body, and what is not part of the body.
But it has access, in this case, to other sorts
of sensory signals, not just visual signals,
or tactile signals, but also proprioceptive signals.
The orange arrows here.
These inform the brain about the body's configuration
and position in space.
And then also, and often overlooked,
are interoceptive signals.
These are signals that originate from within the body, that
tell the brain about the physiological state
or condition of the inside of the internal physiological
milieu.
And you can think the idea is that our experience
of embodied self-hood is the brain's best
guess of the causes of all the signals put together.
Yeah, that's just to emphasise interoception.
An important part of this idea is
that interoception, the sense of the body from within,
should work along the same principles, the same Bayesian
principles that we've been thinking about,
vision and audition, previously.
That is, our experience of the inside of our bodies
is the brain's best guess of the causes of the signals that come
from the inside of our bodies.
So we can think of, again, top-down predictions
carrying predictions about what the bodily state is like,
and bottom-M up prediction errors that
report the differences between what's going on
and what the brain expects.
So what is our experience of the inside of our bodies?
Well, way back at the beginning of psychology,
William James and Carl Langer proposed
that emotions, emotional experience,
was really about the brain's perception of changes
in its physiological state, rather than perception
of the outside world.
So, in this classic example, seeing a bear
does not in itself generate the experience of fear.
Rather seeing the bear sets in train,
a load of physiological changes preparing for fight and flight
responses.
And it's the perception of those bodily changes
in the context with the bear being around that leads
to our experience of fear.
So the Bayesian perspective just generalises that idea,
and says that emotional experience
is the brain's best guess of the causes
of interoceptive signals.
And this fits very nicely with a lot of evidence.
And this is just one study done by a Finnish group.
And all they did here was, they had people report
where on their bodies they felt various emotions to take place.
And so you feel anxiety in one part of your body.
You feel fear in another, and so on.
So our experience of emotion does
seem to be intrinsically embodied.
Now another part of our experience of being a body
is the body as a physical object in the world.
And this might seem quite easy to take for granted,
since our physical body is just always there.
It goes around with us, it changes over the years,
in unfortunate ways.
But it's always there.
But it would be a mistake to take our experience of body
ownership for granted.
And there are some classic experiments
that demonstrate how malleable our experience of body
ownership is.
This is the famous rubber hand experiment.
Probably some of you have seen this.
What happens here is that a volunteer has their hand hidden
under a table, and the fake hand is put on top of the table,
and then both hands are simultaneously
stroked with a paintbrush.
And it turns out that just seeing a hand-like object where
a hand might be, feeling touch, and then seeing
the object being touched, is enough evidence
that the brain's best guess becomes that fake hand is,
in fact, part of my body.
Sort of part of my body.
This is what it looks like in practise.
Here you can see the fake hands, focusing on it.
There's the real hands, not focusing
on it, simultaneous stroking-- and there are
various ways you can test it.
[AUDIENCE LAUGHTER]
- I found in doing this, it works even better on children,
by the way, if you do that.
So that's interesting, because that's
using visual and tactile signals to convince
the brain that this object is part of its body.
In my lab, we've been interested in whether these
signals that come from inside the body also play a role.
So we set up a virtual reality version of this rubber hand
illusion, where people wear these goggles,
and they see a virtual fake hand.
And we also record their heartbeats.
And now what we can do is, we can make the virtual hand
flash either in time or out of time with their heart beat.
And we asked the question, do people
experience this virtual hand as more
belonging to them when it's flashing
in time, rather than out of time, with their heartbeat?
And the answer is that it does.
And this is just some data, basically that,
bigger than that, which means that, indeed, they experience
the hand as more their own.
The way we measure that actually,
is that first we can ask them.
That's the easiest way.
Then we can also ask them to point
to where they think their hand really is,
and we can see how far they drift
from where their hand really is to where the virtual hand is.
And that provides a more objective way
of measuring the strength of the effect.
Here's what it looks like in practise.
Again if you can see this, that's the real hand.
That's somebody's virtual hand.
Again, imagine you're wearing a headset
so you'll see this in 3-D. And you
can just about see it pulsing to read and back, I hope.
And you can also do some other things
with these virtual reality rubber hands
that you couldn't do with real rubber hands.
For instance, you can map movements of the real hand
to the virtual hand, so you can start to ask questions
about the extent to which the virtual hand moves
as I predict it to move.
How much does that affect the extent to which I
feel it to be part of my body?
You can make it change colour.
So you can have somebody embody a skin colour associated with
a cultural out-group, and see if they become less racist
as a result.
And then my favourite is where you can change, actually,
the size of the body.
And that's coming up here.
So here what we do is, we can have the hand telescope
up and down in size.
And again, this might seem like fun, and it is fun,
but there is a serious purpose.
There are various conditions.
There's in fact, a condition called
Alice in Wonderland Syndrome, where people report
that all parts of their body are, indeed, telescoping
up and down in size.
And in a more subtle way, there are lots of body dysmorphias,
of subtle misperceptions of body shape, which
might be associated with eating disorders.
And so these sorts of techniques allow us to approach,
in a very fine grained way, how people might
mis-perceive their own bodies.
That brings me to the third take-home message about self.
And with apologies to Descartes, the take home message
is that, I predict myself, therefore I am.
In the last nine minutes, before the bell rings,
I want to go full circle and return to this Cartesian idea
of the beast machine.
To try to convince you that our experience of being
a conscious self is intimately tied up with our beast machine
nature.
And to do this, I need to mention one final aspects
of perceptual inference, which has a lot
to do with Karl Friston, who's done a lot of work
in the Bayesian brain UCL here in London.
And if we think of the brain as being in the business
of minimising prediction errors, this can be done either
by updating our perceptual predictions,
which is what I've been talking about so far.
And this what Helmholtz said.
Or we can minimise prediction errors by making actions.
We can change what we predict, or we
can make an action so that our predictions come true.
You can change with sensory input,
or you can change what you believe
about your sensory input.
One point of doing this is, that you can make actions,
then, to find out more about the world that surrounds you.
And this is what Helmholtz has in mind
when he says that each movement we make,
by which we alter the appearance of objects,
should be thought of as an experiment designed
to test whether we've understood correctly
the invariant relations of the phenomena before us.
Which Gregory, much later, said something similar
when he talked about perception as hypothesis testing.
The point of this is, that we make
eye movements, and other kinds of movements,
to understand what the world is like.
That, in fact, there was is tomato there, for instance.
But there's another way to think about active inference,
which is that, when we minimise prediction error,
what we're actually doing is controlling a sensory variable.
We're preventing it from changing,
because we're making our prediction about what
it is come true.
And this is the use of active inference,
to control or regulate something,
rather than to understand what the causes of that something
are.
And this brings a very different tradition
to mind, which is 20th century cybernetics.
And this is Ross Ashby, who was a pioneer
of this way of thinking.
And he, with Roger Conan right at the end of his life,
wrote a paper.
The title of the paper, was "Every good regulator
of a system must be a model of that system".
The idea here is, if you want to regulate something very
precisely, then you need a good model
of what effects that system.
Now you could apply this idea to the external world, as well.
When you try to catch a cricket ball,
you are actually trying to control the level
of the angle above the horizon.
But it applies more naturally, I think,
to the internal state of our body.
So really, what matters about my internal physiological
condition, I don't really need to know
exactly what it's like inside my body, and care about it.
But I need to control it, and my brain needs to regulate it.
So this way of thinking about active inference
applies more naturally to interoception.
Think about it this way that.
Having good predictive models are always useful,
but we can have a pendulum that swings, on the one hand,
towards control.
We can use these predictive models for control,
and that's more applied to the state of our internal body.
Or we can swing the other way, and think about perception,
understanding.
You could think of the instrumental, epistemic ways
of thinking about the role of action and perception.
And this brings to mind--
I mentioned Karl Friston.
He's come up with this thing called the free energy
principle.
And I can only nod to the vast body
of work he's done here on this.
With the slogan, which is that organisms, over the long run,
maintain themselves in states in which they expect to be in,
in virtue of having good predictive models
about their own internal condition.
So this takes us right back to Descartes,
but in a very different way.
As I said right at the beginning of this lecture, for Descartes
our physiological reality was rather
irrelevant to our minds, our rationality, our consciousness.
This is a quote from a 1968 paper on his beast machine
argument, "without minds to direct their bodily movements,
animals must be regarded as unthinking, unfeeling machines
that move like clockwork."
Now I think if you try to think how
this idea of our predictive models
controlling our internal physiological states,
and the resulting experiences that perceptual content that
might give rise to, you can make the opposite case.
And the opposite case would be that conscious self-hood
emerges because of, and not in spite of, the fact
that we are beast machines.
And I think this is a deeply embodied view of consciousness
and self, and it speaks to this fundamental link in continuity
between what it is to be alive, what it is to have a mind,
and what it is to be a conscious self.
So I repeat, the third take-home message should
make even more sense now.
That I predict myself, therefore I am.
And I am a conscious self because I'm
a bag of self-fulfilling predictions
about my own physiological persistence over time.
Now why does any of this matter?
It's a lot of interesting ideas, but why should we
be interested in studying consciousness?
Well it's a very interesting thing,
I hope I've convinced you.
But there are lots of practical reasons
to be interested as well.
There are between 20 and 60,000 patients in the UK
alone, who are in disorders of consciousness.
You are in the vegetative state, or in coma,
or in some other severely abnormal state
of consciousness.
Having better measures of conscious level,
as I described at the beginning, is
going to really change again, and how
we treat people like this.
And of course, in psychiatry.
Psychiatric disorders are increasing that prevalence
across all modern societies, and it's
estimated one in six of us, at any one time,
are suffering from a psychiatric condition.
And understanding the mechanisms that underlie conscious
content and conscious self, because a lot
of psychiatric disorders include disturbances
of the way we experience our body,
even though that might not be the most obvious symptom,
can help us understand the mechanisms
involved in psychiatric disorders,
not just the symptoms.
There are also some more general reasons
for studying consciousness, which bring up
some ethical questions.
When does consciousness emerge in development
on newborn babies conscious?
Or does consciousness start even in the womb?
Maybe different dimensions of consciousness
emerge at different times.
Are other animals conscious?
Well I think it can make a very good case for mammals
and primates, but what about the octopus?
The octopus has more neurons in its arms
than in its central brain.
They're very smart creatures.
Here, you have to ask the question not only, what
is it like to be an octopus, but what is
it like to be an octopus arm?
And finally, with the rise of artificial intelligence,
we should begin to ask questions about what would it
take for a machine to have some kind of subjective experience.
I don't think we're anywhere near that yet,
but we should consider what science can tell us
about its possibility, because that would raise some very,
very tricky ethical questions.
But, fundamentally, consciousness
remains fascinating for me for the same reason
that it's motivated people throughout the ages.
I mean, Hippocrates, the founder of modern medicine,
put it one way.
He said, from the brain and from the brain alone arise
our sorrows, our joys.
And he also had his first view of psychiatry
that madness comes from its moistness.
And then Francis Crick, in the 1990s, who
I mentioned in the beginning.
He gave birth, if you like, to the modern neuroscience
of consciousness.
He said much the same thing in his astonishing hypothesis.
But there is this mystery and wonder still,
about how the biological machinery inside our heads
gives rise to the rich inner universe that
makes life worth living.
And despite this mystery, modern science is making progress.
I hope I've given you a flavour, even though we don't understand
how consciousness happens, we can begin to understand its
mechanisms.
So we should not be afraid of naturalising consciousness.
It's not a bad thing to understand its basis
in the material world.
As so often in science, with greater understanding
comes a larger sense of wonder, and a greater realisation
that we are part of, and not apart from, the rest of nature.
[AUDIENCE APPLAUSE]