Mind Reading

Mind Reading


Mind reading? Of course not. I love reading. Look, mind reading might sound
like pseudoscientific– pardon my language– bullshoot. But its scientific counterpart,
thought identification, is very much a real thing. It’s based in neuroimaging
and machine learning, and what’s really cool is
that experiments in mind reading aren’t just about spying
on what someone is thinking. They’re about figuring out
what thoughts are even made of. I mean, when I think
of something, what does that mental picture
actually look like? What resolution is it in? How high fidelity
is a memory, and how do they change
over time? Well, in this episode, I’m going to look at how
reading people’s minds can help us answer
these questions. My journey begins right here
at the University of Oregon. I’m meeting with Dr. Brice Kuhl
from the Kuhl lab. He’s a neuroscientist
who uses neuroimaging and machine learning to figure
out what people are thinking without them telling him. So tell me what
you’re doing here. Well, I’m in the cognitive
neuroscience program here, and I study human memory. My lab primarily uses
neuroimaging methods, so we do a lot of work using functional magnetic
resonance imaging, or fMRI. And how do you use
fMRI to investigate memories? We’re looking at the pattern
of neural activity. When you form a memory,
there’s a certain pattern. And we can record
that pattern and then test whether
that pattern is reinstated or reactivated at a later point,
like when you’re remembering it. Does that mean we can look at
the patterns of brain activity and deduce what it is that is
being remembered, or recalled, or even just thought? Yes, and so we call that
decoding. So it basically takes
your input pattern as some pattern of activity
that we record while you’re remembering
something. And we make a prediction
about what you’re remembering. You can see how this sounds
like mind reading. [laughs]
Yes. It sounds like that. So, Brice, what are you going
to do to me today? So, what we’re going
to be doing today is uncharted territory
for us. So we’re going to be trying out
a kind of new variant of the experiment on you. So I can’t guarantee
any particular results. But it represents
where the field is and where
we’re trying to go. Today, you’re going to
participate in an experiment where you’ll be studying faces. So we’re going
to have you study 12 pictures of celebrities. People I already am
familiar with. -People that you know, yeah.
-Okay. And you’re going to try
to remember those pictures. Then we’re going to have you go
into the MRI scanner. Try to bring that picture
to mind as vividly as possible. And we’re going to be recording
your brain activity as you try to imagine
these pictures. We’re going to try
to build the face. Essentially draw a picture of
what you’re remembering. -A picture?
-A picture. An actual picture
that we can print out and I could, like,
hang on my wall. [laughs]
If you wanted. [Michael]The first step
is for me to memorize
the 12 specific
celebrity photographs
Brice will later try
to detect me thinking about.
I sat down to do this
graduate student, Max.
The success of his predictions
depend, in part,
on my ability
to recall these faces
as vividly as possible
while inside the fMRI.
All right, so… [sighs] I think I have a pretty good
memory of all of those. -Great.
-I feel the stakes are high.With the celebrity faces
hopefully memorized,
it’s time for the next step:going through
the metal detector
and into the fMRI,where Brice will record
and monitor my brain activity,
and then later feed it into his
algorithm to rebuild the faces.
This will be the first time
he’s attempted
to reconstruct faces
from long-term memory,
which is very difficult,
because we’re relying
on how clearly I can remember
the celebrity photos
I saw an hour ago.I love its eyes.
Look at that. [woman] Wouldn’t the kid be like,
“It’s going to eat me”?An fMRI monitors the activity
within the brain
by dividing it up
into thousands of small cubes
called voxels,
or volumetric pixels.
Each of these voxels containshundreds of thousands
of neurons.
Using fMRI,
we are able to detect
blood flow
within these voxels,
which means that that part
of the brain is active.
If I’m shown several pictures
of people with mustaches,
my brain will react
to the features for each face.
But there will be
a common area of my brain
that is engaged
throughout.
That may be the area of my
brain that reacts to mustaches.
So later,
when I imagine a face,
if Brice notices
that area is engaged,
he can predict
that I am thinking
about a mustache.So right now Michael’s
in the scanner, and he’s seeing words appear
on the screen one at a time, and he’s trying
to visualize the face, remember the face in as much
detail as possible. What you can see here are
the images that we’re acquiring. We get one of these
brain volumes every two seconds. So these are refreshing in real
time as we collect the images. [Michael]With part one
of the fMRI session over,
it’s time for part two,
where Brice and his team
will learn the language
of my brain activity,
so they can later
decode by brain scans.
Hi, Michael.
You doing okay still? [Michael]
Yup.They’ll show me hundreds
of unique faces,
and record how my brain reactsto certain facial
characteristics.
They will then use
this information
to reconstruct
the celebrity faces
I thought about during
the first phase of the scan.
Really, the more faces that
we can show Michael, the better. So we’re going to basically keep
him in there as long as he’s comfortable. [Michael]
Two hours was the maximum time
we could get in the fMRI.But I was able to look
at over 400 faces,
which should be enough to getsome pretty interesting
results.
Hey, Michael, you did it.
That was great. We’re going to come
get you out. [Michael]
All right. Yeah, so these just show
some of the pictures that we were taking
while you were in there. Some images of your brain. Now we are going
to crunch some numbers. Max is going to analyze
your data. We’ll meet up
again tomorrow, where we’ll look
at the results, where we try to actually
reconstruct the face images from the brain data
that we just collected. All right.
Well, see you tomorrow. All right.
Thanks a lot. Max, thank you as well.
I can’t wait. You better pull
an all-nighter. I want this data
to be perfect. All right, so I am back
at Dr. Kuhl’s lab. Overnight, his team
crunched the data, and I can’t wait to see what
they think they saw me thinking. How are my results? I think they look good. We’re going to take a look
in just a moment here. All right,
I can’t wait. -So can I just take a seat?
-Yeah, have a seat. All right, so… first of all… what am I seeing?
Oh, okay, well, these are the pictures
I actually memorized. -That’s right.
-And this is what you’ve reconstructed
from my imagination. -That’s right.
-Oh, wow. Okay. [Brice]
Okay, so this is one
of the reconstructions that was generated. [Michael]
Interesting. [Max]
So that’s John Cho. [Michael]
Not bad. Not bad. -Can we see the side by side?
-Yeah. [Michael]
I see, you know, similarities in the kind of facial
expressions in general. You know, you could almost
see the hairline matching here. The shape of the face
I also thought was– It kind of had
a square shape to it. -Yes. Yes.
-So those are the things that came out to me. And so when I was
visualizing this image of John Cho, the squareness of the face was
the first, most salient thing. I just kept thinking,
he was the square guy. Excellent, all right. [Brice]
So that’s Megan Fox. [Michael]
Mm-hmm. You’re going to show us the–
side by side. [Michael]
The side by side. Right. [Brice]
You can see the picture
you actually saw, and that’s the reconstruction
we generated. I’ll you this.
Megan Fox, I was not able to have a really clear picture
in my mind. For some reason, this image
of her was really hard for me to bring back into my mind. The sternness in the face was
something that I did pick up on. So I did sense that there was–
It looked feminine. And you picked up
on the sternness. And so together,
that produces a match. [Michael]Keep in mind
that Brice and his team
have read these
from my memory.
But when I remember a face,do I picture every detail
simultaneously
with photographic accuracy?Or do I just attend to a few
at a time?
By reading my mind,
they may be seeing
how bad my memory is,
and how it works.
-Me! Me!
-[Brice laughs] Okay, so that is your
reconstruction of me thinking about
this image of myself. [Brice]
That’s right. Where’d the beard go? [Brice] I don’t know.
I was hoping you could tell me. [Michael]
For instance, this is a picture
of me remembering my own face.
It really doesn’t look like me,
but the question is:
how good am I
at picturing myself?
I don’t think of my own face
that often,
so the strangeness
in the result
may be as much about flaws
in my own memory
and mental picture of myself
as flaws in the technology.
So that’s Jennifer Lawrence,
I believe. [Michael]
That’s Jennifer Lawrence? It looks like it’s Jennifer
Lawrence’s much older uncle. [all chuckle] Nothing here was too
mind-blowingly close. But this is something that
you’re just starting out trying these sort of long-term
memories.What Brice and his team
read in my mind
might have been more accurate
if they’d shown me thousands
rather than hundreds of images
in the fMRI,
because then the algorithm
would have learned
the language of my brain
more thoroughly.
But regardless,
the quality of my memories
would have still
been an issue.
I mean, look what happens
when memory
is cut out of the equation
entirely.
Brice also read
my brain activity
when I was looking
at faces in the fMRI.
not just imagining them.And those results
were much closer
than those reconstructed
from my memory.
Okay, so, what am I looking at
right here? [Brice]
Okay, so what you’re seeing here in the top row,
these are images that you saw while you were in the scanner. Below that, in this bottom row,
these are the reconstructions that we draw from the patterns
of brain activity we collected. -This is from the source image.
-Right. [Michael]
These are from my brain. -[Brice] Right.
-[Michael] They’re pretty close. Yeah, overall they were
pretty close. So not perfect. These are– you can see there’s
some variability in these. But this is consistent
with what we’ve found before, that the reconstructions
that we generated, when you’re viewing
the faces, there is some correspondence
between the actual face. So this is kind of
a sanity check, that we can actually
reconstruct the images -when you’re viewing them.
-Right, right. They’re pretty good. Well, Brice, Max,
thank you so much for letting me
be a part of this. I hope my data’s useful. Thank you.
It’s been a lot of fun. It’s always useful for us
to think about these things. Dr. Brice Kuhl’s memory research
is showing that it’s possible for a computer
to read someone’s mind. To figure out
what they’re thinking. But a lot of progress
still needs to be made. I mean, if you want to know what I’m thinking right now,
for example, it’s still easier to just ask me
to tell you. But what if I can’t
tell you? Dr. Yukiyasu Kamitani
is a researcher, professor and pioneer
exploring the frontier behind the wall of sleep. I’ve come here
to Kyoto University to meet with him and to see
what it’s like to read not what someone
is thinking, but what someone
is dreaming. Kamitani sensei,
I’m Michael. -Hi, I’m Yuki.
-Yuki, nice to meet you. [Michael]
For the last ten years,
Dr. Kamitani has been
at the forefront
of machine mind reading.The subject is, you know,
ready to go in.Similar to Brice Kuhl,his early experiments explored
reconstructing images
shown to subjects in an fMRI
based on their brain activity.
In Kamitani’s case,the images were
black-and-white shapes,
and the reconstructions
were strikingly accurate.
Recently, Kamitani has focused
on using deep neural networks
and machine learningto decipher subjects’
brain activity
while they view
much more complex photographs.
What you’re seeing is the
result of a deep neural network
processing the brain activity
of a subject
looking at the photograph.This could have myriad
applications in the future,
for example,
in criminal investigations
and interpersonal
communication.
[Kamitani]
This is far from perfect. But I think you still see some,
you know, eyes and, you know… [Michael]
Well, yeah. And colors too. [Kamitani]
Yeah, to some extent, yeah.His most current work, however,
is about the subconscious.
He’s attempting something
extremely ambitious:
recording our dreams.Would you call yourself
a sleep researcher, or a vision researcher? Maybe a brain decoder. A brain decoder. That’s a pretty cool
job description. Can you show me anything from
what you’re doing with dreams? [Kamitani]
Mm-hmm, yeah.Dr. Kamitani’s work
on dream decoding
begins with a similar process
to Dr. Kuhl’s:
showing the test subject
thousands of images
while they are in an fMRIin order to learn
what the brain looks like
when it is thinking
of certain things.
Once the machine-learning
algorithm is pretty good
at identifying what images
the subject is thinking about,
the subject is placed
in an fMRI
with an EEG cap
on their head,
and invited to fall asleep.When the EEG waves indicate
that the person is dreaming,
the algorithm predicts
which kinds of things
the subject is most likely
dreaming about.
Right now, the algorithm
looks for 20 categories.
Things like buildings,
transportation,
and characters
in a language.
Researchers then awaken
the subject,
ask them what they were
dreaming about,
and see if the algorithm’s
prediction
and the person’s
recollection match.
Here is actual data from
one of Kamitani’s experiments.
Below is a word cloud
of categories.
The name of each category
get bigger or smaller
in real time
based on the probability
that they are present
in the subject’s current dream.
Now, as you can see,
activity is currently strongest
for the category “character,”
meaning written language.
At this point
the subject was awoken,
and this is what
they reported.
That’s pretty spooky. -[laughs]
-Right? I mean, you– you spied on their dream. Yeah, in a way.
But… the accuracy’s
not that great, so… Well, the accuracy’s
not that great but, you know, my normal accuracy for guessing
people’s dreams is zero. Right.While continuing
his research
into predicting
the content of dreams,
Dr. Kamitani is embarking
on his newest project:
actually reconstructing images
from our dreams.
So you’ve brought
some of the reconstructions that your lab has created… Mm-hmm. …of dreams. Right, they all look like dreams
about blobs. [Kamitani]
Yeah. I mean, I want to just
take a step back and… appreciate that what we’re
looking at on this screen are, in a way, some of the first
photographs of a dream. Mm-hmm.We are looking
at the earliest phase
of revolutionary research.One day, we may able
to have images,
or even record movies,
of our own dreams.
And Dr. Kamitani is the only
person in the world
doing this so far.He’s a lone explorer journeying
into our subconscious.
So this work hasn’t even
been published yet. No. -Thank you for showing it to me.
-[laughs] The insights that researchers
like Dr. Kuhl and Dr. Kamitani might be capable of achieving
in the future because of mind reading are difficult
to fully fathom. But let’s slow down
for a second, because we’re talking
about a technology that can know us
better than we know ourselves. Should we really
be doing this? Well, to address
that question, I’m going to meet
with an expert in ethics, neuroscience
and artificial intelligence: Julia Bossmann. She’s the director of strategy
at Fathom Computing, a council member
of the World Economic Forum, an alum of Ray Kurzweil’s
Singularity University, and a former president
of the Foresight Institute, a think tank specializing
in future technologies and their impacts. Julia, thanks for taking some
time to chat. -Yeah, of course.
-You are the perfect person for me to bring
these questions to. -Mm-hmm.
-And they’re deep questions. But I think they’re
extremely important, and they’re becoming
more and more pressing. I think we’re living in such
an interesting time right now, because we’re in this time
where brains and machines are actually moving
closer together. So when it comes
to being able to look at brain activity, where are
the ethical lines here? How private should
my internal thoughts be? Like with any powerful
technology, it depends on the hands
that wield it. All these new technologies are things that can make whoever
uses them more powerful. So we want to not blame
the technology, but we want to– how is it being used, and who is using it? So how do we make sure
that this technology is in the right hands? So I think it’s very important
to involve people who act on policy and law to understand what is coming
in the future. I am hopeful about
the collaborative aspect of it. Let’s talk about
the good things now. I mean, what are
the applications here? Yeah, so if we think about the late Stephen Hawking,
for example, if he had a way of richer
interfacing with the world or with computers,
we can only imagine what he could have
shared with us. Those with locked-in syndrome,
right? They are there.
They know that they are there. But we just need something
to look into their brain to see what it is
that they are trying to say, -or what they’re feeling.
-Right, exactly. So, what do you say
to people that have that kind of fear
of technology, of us surrendering our true
natural selves to technology? There is something enticing
about getting to the next level of what some people might call
a human evolution or civilization development,
and so on. In a way, we are already not
living natural lives, right? Because then most of us would
die before the age of, I don’t know, 30 or 40. We would have all kinds
of diseases. We would not wear
this clothing. We wouldn’t have eyeglasses
or contact lenses. We wouldn’t have antibiotics. [Julia]
We are already kind of very futuristic cyborgs
if we compare ourselves to the human that was living
10,000 years ago and was genetically
almost identical with who we are now. [Michael]
Yeah, we really are. In order to understand
cognition, right now we basically have
to either just ask people to talk about
what they’re thinking, or observe their behavior. But reading thoughts directly
would be a lot better. That is how Dr. Kuhl
is studying memory, and it’s how Dr. Kamitani
is studying sleep and dreams. But even though the technology
has a long way to go, it’s easy to see
how ethical questions could become an issue. Well, here’s the thing: there is no such thing
as a totally wild human. We are co-evolving
with technology. Humans and technology today
are inseparable. Now, it’s true that we need
to be careful about every new thing we do, but we cannot change the fact
that they will happen. It’s a story we’ve lived through
again and again. You know, we could have
sat around forever debating whether or not
a speed limit should exist and who should have
the authority to enforce it. But we didn’t. Instead, we went ahead
and invented cars, and responsibly figured out
the details as we went along. Ethical questions
about new technologies do the most good when
they facilitate the technology, not when they needlessly
hinder progress. So follow your dreams. And, as soon as you can,
show them to me. And, as always,
thanks for watching.

65 thoughts on “Mind Reading

  1. Hablas de tecnología. ¿Sabes lo que es realmente tecnología? Es el conocimiento de una técnica. Por lo tanto, una casa, un fósforo o incluso un corta uñas es tecnología. Está mal explicado el térmico, Michael. Hablas de tecnología como si fuera algo SOLO de electrónica o algo más allá, cuando es algo mucho más simple. Por favor, explica en un video de Vsauce lo que REALMENTE es TECNOLOGÍA. Saudos.

  2. Just imagine in a couple of years how this could affect the lives of nonverbal people who can’t use current speech tools or people in a vegetative state, this is amazing work

  3. what i understand: Michael see every imperfection in his beard and he's automaticaly imagining a grey part in beard place for every face

  4. I had a dream before that I was hooked up to a projector and it was playing my old dreams. I guess that’s becoming a real thing………..

  5. Dr. Kami is low key flexing on all of us cause he just out here reading people’s minds and he’s just like “well, it’s not THAT accurate…”

  6. "Ethical questions about new technologies do the most good when they facilitate the technology, not when they needlessly hinder progress."
    Excellent point to jot down. Love it. Thanks for making these available for all to watch Michael – that's super cool of you! I signed up for YouTube Red I think its called when these first came out so I could pay to watch them – not sure how much of that goes into your pocket, but nonetheless want you to know I appreciate your videos and wanted to spend the money I'd worked for on hearing what you've got to say.

  7. "or even record your dreams"
    first thing that comes to my mind:
    "iMaGiNe ReCoRdiNG yOuR eRoTiC dReAmS jUsT tO LaTeR wAtCh YoUrSeLf FuCkiNg PeOpLe YoU cOuLdNt IRL"

  8. Normal brains: ok so tonight we’re going to dream about eating your best friend’s cat as he eats yours, and it’s a ritual of friendship, but you’ll both pretend it’s completely normal until the elephant knocks on your door to deliver your mail
    That one guy: I dreamed of writing some sort of essay or filling out a form

  9. I love this! Super cool to know that people are beginning to get a glimpse of how to interface the human brain so we can get to see it from outside. So super cool! There might be regulations we have to put along the way, it could be scary if used wrongly. But i think we should embrace this change, along with many others that are on their way

  10. Imagine what quamtum computers will be able to do once theyre fully funtional.

    Its amazing and scary at the same time.

    The machines are about to wake up and record our dreams….

  11. Just About To Watch The Vid, So I thought I would pass my own thoughts on regarding this subject, to use a scientific analogy – two large old style brick monitors in different rooms one is hooked up to a tuner card, by tuning the frequency to the monitor in the other room the same picture can be seen on the monitor with the tuner card Therefore since all human beings have the same equipment a mind then the Mind can tune into the frequencies around them an "MIND RECEIVE" impressions and thoughts, "Hey, I Was Just Thinking That" – People automatically synchronize with those around them as seen with women and their time per month happening at the same time. There is also the concept that all minds on the planet are on a subconscious level connected mentally and that the six degrees of separation allows all people to be connected in a huge pool of super consciousness from which we through our "Creativity" are able to tune into this "Giant Pool Of Networked Thought" to gain insights and muse, while still retaining our own individuality, the direct link to that pool may have one been active however was broken in evolution (Tower Of Babel) as we gained individuality and personal choice….. Just My Thoughts Before Watching The Vid

  12. 1:48 The Question comes are the patterns specific to an individual or are their common elements in which case if a person was learning physics could that's persons commonality then be used to implant those patterns in another person brain for a "Learning Machine" like "L Ron Hubbards Battlefield Earth Book And Movie Learning Machine" could those same patterns also be used to create a Personal AI like Cortana, Or Better Still One Of The Tele-Presence Robots With An MRI Of Those We Love So That They Can Still Pass On Their Wisdom To Us Once They Have Passed (Independence Day Movie Sphere's Virtual People) Or Supermans Holographic Dad ..(Lemurian Crystals – The Deep Season 01 Trailer)

  13. 9:15 Get A Pic Of When You Were Young And Test Resemblance From Back Then, When Did You Like The Way You Looked In The Past The Most!

  14. 21:02 Who Get The Biscuit First! 21:44 Maybe! – Use Satellites that land on comets and meteorites to alter their course (Ion Drives) and create a couple of new planets in farther orbits around the sun, hoping that there is large chunks of ice floating around there (Comet Watch) so we can make them habitable, and maybe use a couple to change Venus from Co2 To O2 & H2O NB:Pity We Lost The Big Cigar Comet would have provided a good Iron Ferrite Core

  15. that last hit about ethical debate hindering progress- awesome. someone had to voice that. i can't stand when people dive into ethical debate on things when you can't even be sure what would be ethical or not until you actually go through with the idea(s). we can figure it out along the way. trial and error is the only way to move forward. there's no way to learn to ride a bike without taking a tumble once or twice.

  16. Your memory of your face may be an amalgamation of all instances of "your face" committed to memory as you age.

  17. I just imagine them doing a dream experiment and then ask the person what he dreamed of and they're like… i don't remember.

  18. the fears concerns only our individual lives of today… what should happen to me, or to my girlfriend, if…
    but there are lots of subjects to fear. The death is one other.
    But the humanity has NOT to be exactly as the humanity of 6000 years ago, and same thing for birds. All ways of lives will surely co-exist in the future, as nowadays. Travelling in forest countries and we see peaple living nudes and hunting.
    ( despite the problem of deforestation ! but is NOT a new-technology problem )
    So, maybe will be there "cyborgs" humains… more far from natural life… but close from new realities, and…
    i "believe" ( = have the hypothese ) that "universe wants to do always more experiments".

    . . . because it is needed that there is << "something" instead of "nothing" >>, and the "something" must be more and more complex to exist…

  19. This was by far one of my favorite episode. I enjoy the peeks into current research much more than I enjoy the social experiments on mind field.

  20. Si el cluster de neuronas fuera mas preciso por ejemplo docenas en vez de miles de neuronas el resultado de la reconstrucción seria mas preciso!! De todas maneras el estudio de kamitani esta mal, deberia mostrarle imagenes menos abstractas durante el analisis a los sujetos, por ejemplo pixeles o gradientes de colores en la pantalla, para luego poder reconstruir una verdadera imagen, no un promedio de significados entre imagenes, por eso siempre obtiene una esfera de colores y no una imagen

  21. I’m so interested with this kind of thing and want to learn more but at the same time I feel that dreams are sacred and shouldn’t be messed with

Leave a Reply

Your email address will not be published. Required fields are marked *