Blaise Agüera y Arcas: How computers are learning to be creative
Blaise Agüera y Arcas works on machine learning at Google. Previously a Distinguished Engineer at Microsoft, he has worked on augmented reality, mapping, wearable computing and natural user interfaces. Full bio
Double-click the English transcript below to play the video.
that works on machine intelligence;
of making computers and devices
that brains do.
interested in real brains
in the things that our brains do
to the performance of computers.
has been perception,
out there in the world --
for example, that our team makes,
on Google Photos to become searchable,
out there into the world.
our work on machine perception
with the world of machine creativity
had a penetrating insight
between perception and creativity.
has a statue inside of it,
is to discover it."
Michelangelo was getting at
is an act of imagination
and perceiving and imagining,
with a brief bit of history
the heart or the intestines,
about a brain by just looking at it,
of this thing all kinds of fanciful names,
doesn't tell us very much
developed some kind of insight
Santiago Ramón y Cajal,
or render in very high contrast
their morphologies.
that he made of neurons
of different sorts of cells,
was quite new at this point.
very, very long distances --
to some people in the 19th century;
were just getting underway.
of Ramón y Cajal's, like this one,
that Ramón y Cajal started.
of Neuroscience.
is about one cubic millimeter in size,
very small piece of it here.
tiny block of tissue.
of hair is about 100 microns.
much, much smaller
electron microscopy slices,
in 3D of neurons that look like these.
style as Ramón y Cajal.
be able to see anything here.
one neuron to another.
ahead of his time,
over the next few decades.
was advanced enough
experiments on live neurons
when computers were being invented,
of modeling the brain --
as Alan Turing called it,
looked at Ramón y Cajal's drawing
imagery that comes from the eye.
like a circuit diagram.
in McCulloch and Pitts's circuit diagram
of computational elements
one to the next in a cascade,
visual information would need to do.
for us to do with our brains.
that for a computer,
just a few years ago.
this task is easy to do.
and the word "bird,"
connected to each other
inside our visual cortices,
to have the capability
on the computer.
that actually looks like.
about as a first layer of neurons,
how it works in the eye --
after another layer of neurons,
of different weights.
of all of those synapses.
properties of this network.
or a small group of neurons
those three things --
in the neural network,
these synapses in the neural network.
is just a simple formula,
going on there, of course,
of mathematical operations.
that if you have one equation,
by knowing the other two things.
that the picture of a bird is a bird,
and w and x are known.
you know the pixels.
a relatively straightforward problem.
and you're done.
doing exactly that.
on a mobile phone,
amazing in its own right,
billions and trillions of operations
picture of a bird,
"Yes, it's a bird,"
with a network of this sort.
and the y is the unknown.
difficult part, of course,
do we figure out the w,
of solving for w,
with the simple equation
it's the inverse to multiplication,
very non-linear operation;
to solve the equation
is fairly straightforward.
a little algebra trick,
to the right-hand side of the equation.
about it as an error.
for w the right way,
to minimize the error,
computers are very good at.
sort of play Marco Polo,
successive approximations to w.
but after about a dozen steps,
which is close enough.
a lot of known x's and known y's
through an iterative process.
that we do our own learning.
this is not a bird."
for those neural connections.
x and w fixed to solve for y;
Alex Mordvintsev, on our team,
with what happens if we try solving for x,
that you've trained on birds,
the same error-minimization procedure,
trained to recognize birds,
generated entirely by a neural network
rather than solving for y,
by Mike Tyka in our group,
of William Kentridge's artworks,
over the space of different animals,
to recognize and distinguish
morph from one animal to another.
have tried reducing
out of the space of all things
over that entire surface,
you make a kind of map --
the network knows how to recognize.
"armadillo" is right in that spot.
of networks as well.
to recognize faces,
in a y that says, "me,"
psychedelic picture of me
multiple points of view at once
to get rid of the ambiguity
or another pose,
another kind of lighting.
this sort of reconstruction,
of different points of view,
his own face as a guide image
to reconstruct my own face.
that optimization process.
more like a coherent face,
with a blank canvas
that is itself already some other image.
that is designed to categorize
man-made structures, animals ...
with just a picture of clouds,
what it sees in the clouds.
you spend looking at this,
will see in the clouds.
to hallucinate into this,
zooms hallucinates, zooms.
of the network, I suppose,
is eating its own tail.
What do I think I see next?"
called "Higher Education" --
marijuana was legalized.
is not constrained.
because they're really fun to look at.
a camera that takes a picture,
writes a poem using neural networks,
has been trained
are very intimately connected.
things in the world,
Michelangelo really did see
any being, any alien
perceptual acts of that sort
machinery that's used in both cases.
and creativity are by no means
that can do exactly these sorts of things.
the brain is computational.
in designing intelligent machinery.
of those early pioneers,
is not just about accounting
we modeled them after our minds.
to understand our own minds better
ABOUT THE SPEAKER
Blaise Agüera y Arcas - Software architectBlaise Agüera y Arcas works on machine learning at Google. Previously a Distinguished Engineer at Microsoft, he has worked on augmented reality, mapping, wearable computing and natural user interfaces.
Why you should listen
Blaise Agüera y Arcas is principal scientist at Google, where he leads a team working on machine intelligence for mobile devices. His group works extensively with deep neural nets for machine perception and distributed learning, and it also investigates so-called "connectomics" research, assessing maps of connections within the brain.
Agüera y Arcas' background is as multidimensional as the visions he helps create. In the 1990s, he authored patents on both video compression and 3D visualization techniques, and in 2001, he made an influential computational discovery that cast doubt on Gutenberg's role as the father of movable type.
He also created Seadragon (acquired by Microsoft in 2006), the visualization technology that gives Photosynth its amazingly smooth digital rendering and zoom capabilities. Photosynth itself is a vastly powerful piece of software capable of taking a wide variety of images, analyzing them for similarities, and grafting them together into an interactive three-dimensional space. This seamless patchwork of images can be viewed via multiple angles and magnifications, allowing us to look around corners or “fly” in for a (much) closer look. Simply put, it could utterly transform the way we experience digital images.
He joined Microsoft when Seadragon was acquired by Live Labs in 2006. Shortly after the acquisition of Seadragon, Agüera y Arcas directed his team in a collaboration with Microsoft Research and the University of Washington, leading to the first public previews of Photosynth several months later. His TED Talk on Seadragon and Photosynth in 2007 is rated one of TED's "most jaw-dropping." He returned to TED in 2010 to demo Bing’s augmented reality maps.
Fun fact: According to the author, Agüera y Arcas is the inspiration for the character Elgin in the 2012 best-selling novel Where'd You Go, Bernadette?
Blaise Agüera y Arcas | Speaker | TED.com