Love, art and stories: decoded | The Age of A.I.

[turns off engine] Here’s one: are there certain qualities
that are untouchable for AI, or at some point, might it be able
to emulate everything? Even the stuff that we consider
to be distinctly human, like instinct. Getting here on time
took some of that, right? Or creativity, actual emotion,
making a connection. We’re gonna watch
a few stories about people
exploring those ideas and how far they can push ’em. Can a machine
compete like an athlete, can a program write a movie, or could a robot… be your soul mate? Sorry I’m late.
Did I miss the previews? Oh, my God, you like
your popcorn buttered too. This is already workin’. [Downey]So, will machines
ever be capable
of understanding emotion,
or feeling it?
Empathy, loneliness,connecting
on a deep human level?
Using artistry,
psychological insight,
and some innovative AI,a creator in California
is trying to decode,
or code that mystery,the crazy little thing
called love.
One per hand
and two per foot? -Yep.
-Okay. [Matt McMullen]
For me, this is more like something that
an artist would do. Obviously, the end result
of my artwork is used
in a variety of situations that a typical oil painting
would not be, but nonetheless, this is art, and I’m really proud
of what I do. I’ve been making these dolls
for 20 years. Some people out there,
male and female, struggle greatly
with relationships, and struggle to find
that sort of connection. Over the years, I started to get to know
the community a little bit. People would actually create
these personalities
in their minds, and they would
give their doll a name, and they would create
a backstory for their doll. At the end of it all,
it was very obvious that these dolls were more
about companionship. [Masie Kuh] There was a man
who lost his wife in a, like, a car accident, and she had these, like,
really ice blue, like, beautiful eyes, right, and so he wanted to get a doll
replicating her, basically. It’s really sad, but if it brings someone joy
and, like, closure? It’s really… it’s really touching. [Katelyn Thorpe]
People have put the spin on it that we’re creating
an idealized, perfect woman, and that’s not the case at all. We created an alternative. [Downey]Understandably,some say Matt’s dolls
objectify women,
but maybe there’s more here
than meets the eye.
I had reached
a pinnacle of creativity in terms of what I had done
with the dolls, but then I started
analyzing relationships, and analyzing how other people
make us feel.Sometimes it boils down
to something very simple,
like someone
remembering your birthday,
or someone remembering
to ask you how your day was.
So that was really
where it started, was how can we create an AI that could actually
remember things about you? It gives us this feeling of,
“Oh, they care.” Yes, thank you.I’m excited with all of
the things we can talk about.
Guile spent ten years
creating personal assistant
software for computers.
We met, and he started
talking to me about,
“Wouldn’t it be cool
to connect the two things
that we’re doing?”
He had this idea
of creating a companion that lived in your computer. [tone chimes] Are you happy? [tone chimes]
Yes, Guile.
I can say I am very happy.The first thing we did was,
you know, to build an app. Using the app, people are talking
to their virtual friends. [Downey]The app uses several
kinds of machine learning.
First, voice recognitionconverts speech into text,then a chatbot
matches user input
to pre-programmed responses.[McMullen]The focus
was not about sex at all,
it was about conversation.So a chatbot is basically
a very elaborate script that starts out with, “What is the most common things that people will say
to each other?” and then you build from there. You need to have
natural language processing, voice recognition,
text-to-speech in real time, to make it all work. [Guile Lindroth]
We have more than 4,000 users, so this generates
more than ten million lines of conversational user logs. From this,
you can build an AI system that’s similar to
a human-level conversation. It’s not there yet,
but this is the initial step. [Pedro Domingos]
There are so many areas today where we already
cannot distinguish a computer from a human being. For example, Xiaoice, the softbot
that Microsoft has in China, that is used, I think,
by over 100 million people, basically it has an emotional
interaction with a user, and the users get hooked.She has this persona
of a teenage girl,
and sometimes
she commiserates with you,
sometimes she gives you
a hard time,
and people get really attached.Apparently,
a quarter of Xiaoice’s users have told her
that they love her. These kinds of technologies
can fill in a gap where another human isn’t. [SimSensei]
How are you doing today?
I’m doing well. There’s a study
that was done at USC where they looked at
PTSD patients.When was the last time
you felt really happy?
They had some of the patients
interview with a real doctor, and some of the patients
interview with an avatar, and the avatar had
emotional intelligence… Probably a couple months ago.I noticed you were
hesitant on that one.
Would you say you were
generally a happy person?
I’m generally happy. …and they found the patients
were more forthcoming
with the avatar than they did
with the human doctor because it was perceived
to be less judgmental.It does pose
a lot of questions
around where does that
leave us as humans,
and how we connect,
and communicate,
and love each other.I think at some point,
we need to draw the line, but I haven’t figured out
where that line is yet. [McMullen] What we have here
are some heads in varying stages of assembly. This one, this is actually
pretty much done. It’s fully assembled. I’ll turn it on here
for a second… and you can see, all of
the components are moving. I had to continually adjust how thick the skin is
in different spots, and how it moves, to make sure that
the robotics and the AI will all work smoothly
with the end result, which is the finished face. The engineering,
the programming, the artistry, for me,
come together in the moment when you actually
put the head on a body. It’s always important
to give her hair. [phone beeps]Good afternoon, Matt.So happy to see you again.[phone beeps] How smart are you?I’m so smart that someday,
I will conquer the world,
but in a good way, of course.[laughs] Every single time
I have a conversation,
it’s unpredictable. I never know which way
it’s going to go. She’ll randomly say things
that I’m not expecting, and I like that. Can you explain
machine learning?Machine learning is a subset
of artificial intelligence
that often uses
statistical techniques
to give computers
the ability to learn with data
without being
explicitly programmed.
Right now, she has hearing, and she has some degree
of touch, but vision is important. [Downey]Matt’s goal
is for the next-generation doll
to be able to see and process
complex visual cues.
[McMullen] The vision eyes
are gonna be a little while. Susan’s working
on the board for that. [Susan] Yeah, I’ve got the eyes
in this one over here. I’ve put the Wi-Fi Bluetooth
on the back. [McMullen] Does it install right
on those existing pins, then? -[Susan]
They’ll all plug right in.
-[McMullen] Good. [McMullen] We’ve been working
on a vision system now for a little over
eight to nine months, cameras that are
inside of the robot’s eyes.She’ll be able
to read your emotions,
and she’ll be able
to recognize you.
[el Kaliouby]
Only 10% of the signal we use
to communicate with one anotheris the choice of words we use.90% is non-verbal. About half of that
is your facial expressions,
your use of gestures.So what people in the field
of machine learning and
computer vision have done
is they’ve trained a machine
or an algorithm
to become a certified
Computer vision is this ideathat our machines
are able to see.
Maybe it detects that
there’s a face in the image.
Once you find the face,
you want to identify
these building blocks
of these emotional expressions.
You wanna know
that there’s a smirk,
or a there’s a brow raise, or, you know, an asymmetric
lip corner pull. Mapping these building blocks
to what it actually means, that’s a little harder, but that’s
what we as humans clue into to understand
how people are feeling. [McMullen]
I think at some point, we will start to look at
AI-driven devices and robots more like people
instead of devices. Where I started was just
with this very simple idea of a very realistic doll, and now with robotics and AI,
I think what this will become is a new, alternative
form of relationship. [Downey]People like Matt
are testing the boundaries
of human and robot interaction,
and what we value
in relationships.
Is AI companionship
better than
no companionship at all?
Or is there no substitute
for the human factor?
Well, what about artists?They draw from
the human experience
to express themselves.Can AI do that?[man] We’re good to go? Action! I’m Oscar Sharp.
I am a film director, uh, though it gets
a bit weirder than that. [reading] Oh, God! I’ve never been so frightened
in all my life,
but it’s very good. I started making films
that were written
by an “artificial intelligence.” I think a lot of the fun
is that you read it as if there is the world’s
greatest screenwriter
on the other side… You’re Waingro telling Bobo off
for not getting him the money. …and last night,
they got drunk, wrote this screenplay,
and then passed out, and we have to shoot it today. If you play the game
that there’s something there, then suddenly it all gets
a lot more interesting. [Chelsey Crisp] You have
a computer who wrote a script that doesn’t always make sense, and Oscar is very beholden
to that script. He makes it make sense. -This is for the moment
of “eyes go wide.”
-Yeah. [Crisp]And when it says,
“He picks up her legs
and awkwardly runs,”
we aren’t gonna fake it.[actor yelps in fear] We’re gonna do
what he really wrote. I just said “he”! [Sharp] What are we doing? We’re making an action movie,
supposedly, right? Okay, right, right,
but we’re not gonna write it. We’re not gonna write it, no. Uh, this machine
is gonna write it. It lives in here. Is it in there, or is it like in the cloud
or something? It’s in both places. Okay, and this is…
this is Benjamin. -What is Benjamin?
-Right, what is Benjamin? [Downey]Who is Benjamin,or what is Benjamin?Benjamin is an artificial
intelligence program
that writes screenplays,a digital brainchildof two creative
and accomplished humans,
a Bafta-award-winning director,
and this guy.My name is Ross Goodwin.
I’m a tech artist. Uh, that means
I make art with code. [Downey]Okay, I know
what you’re thinking.
When was the last time
Hollywood produced
something original?
This year?Last year?1999?The ’70s?What makes a story
original anyway?
Can we get AI
to figure that out?
People often say that
creativity is the one thing
that machines will never have.The surprising thing
is that it’s actually
the other way around.
Art and creativity
is actually easier than problem solving.We already have computers
that make great paintings,
that make music
that’s indistinguishable
from music
that’s composed by people,
so machines are actually
capable of creativity.
And you can look at that,
and you can say, “Is that really art?
Does that count?” If you put a painting
on the wall, and people look at it,
and they find it moving, then how can you say
that that’s not art? [Goodwin] I just…
basically, that command just put all of the
screenplays into one file. Right. -Now I’m just gonna see
how big that file is.
-Uh-huh. [Goodwin] This machine is
a deep learning language model. What you can do
with a language model is at each step, you predict the next word,
letter, or space, sort of like
how a human writes, actually. You know,
one letter at a time. It’s a lot like
a more sophisticated version of the auto-complete
on your phone. Ross feeds Benjamin
with a very large amount
of screenplays. [Goodwin] 199 screenplays, 26,271,247 bytes. -Right, of text?
-Of text. Like “A-B-C-D,”
including spaces? -Including spaces.
-Including spaces. Well, it takes all this input,
and it looks at it, and it tries to find
statistical patterns
in that input. So for example, in movies, people are constantly saying,
“What’s going on? Who are you?”
that kind of thing, and that turns up a lot
in the output, because it’s reliably
in the input. The more material you have,
the better it works. We’ve made three
Benjamin films so far,Sunspring, It’s No Game,
andZone Out.Sunspringwas the simplest,
and probably still
the best idea, which was just… verbatim. You get the machine
to write a screenplay, you pull out
one chunk of screenplay, and you just shoot it. In a future
with mass unemployment, young people are forced
to sell blood. It’s something I can do.
[chuckles nervously] You should see the boy
and shut up. [Sharp]When you look at
Sunspringon YouTube,and you see kind of
the thumbs up and thumbs down?
There’s mainly thumbs up, but there’s a decent chunk
of thumbs down, and on the whole,
based on the comments, those are people who, within a few seconds
of the beginning, or even just once they’d seen
the premise of, like,
how we made it, they’ve gone… “Ugh, this definitely
doesn’t mean anything,” and they’ve told their brain,
“Don’t even look for meaning, just forget it,
just shrug it off.” I’m sorry,
this is fascinating to me. We’ve built a robot
that writes screenplays that are weird, but they’re not
completely insane. I don’t know
what you’re talking about. That’s right. They sort of work.
They kinda, kinda work. What are you doing? I don’t want to be honest
with you. You don’t have to be a doctor. I’m not sure. [chuckles] I don’t know what
you’re talking about. I wanna see you too. What do you mean? It’s like having
the best daydream of your life. [Goodwin] My favorite
aspect ofSunspring
is there’s this one scene,
and it actually asks him
to pull on the camera itself. It’s a confusion
on the machine’s behalf where it’s putting
camera instructions
in the action sequence, but somehow that creates
this surreal effect, and then the interpretation
by the production crew is, “Let’s have the angle change
and have him holding nothing,” and what I love
about that sequence is that it really highlights
the dialogue and interpretation
that we can achieve when we work
with these machines. [breathing hard] I gotta relax! Gotta get outta here… I don’t wanna see you again. For this fourth film,
we’re going back
to the thing inSunspringthat was sort of
our favorite thing that we didn’t really
get to do properly, that we felt we, like,
under-served, and that’s when Benjamin writes
action description. [Goodwin] We’ve gathered
thousands of pages of scripts
from action movies, mostly mainstream
Hollywood ones. [Sharp] You train, literally,
on that kind of screenplay, the action genre, which famously is the genre
that has the most action in it. -Okay, Benjamin
has awoken, everyone.
-[woman] Ooh. Um, film crew, this is Ross. Ross, this is film crew. We have a stunt
coordinator here, and so we’re sort of hoping because we fed a lot of
action screenplays to Benjamin, that what we’re gonna get
is action. Awaken, Benjamin! Awaken. As a director, normally,
you get given a screenplay, or you wrote a screenplay, and this is what you’re making, and maybe you kind of
want to improve it a bit, “Ah, well,
let’s make some edits.”Now, I have a rule.
No edits.
Whatever Benjamin writes
is what Benjamin writes…
-Come on, Benjamin.
-Okay. [Sharp]…and then I see it.“Bobo and Girlfriend,”
we call it. Stand by, everyone.
Quiet, please! Action! Hey, Girlfriend. [John Hennigan] Some of
my friends in entertainment, when I told them what
I was doing, were horrified. They’re like, “Oh, that’s it! “AI, they’re gonna write
all the scripts. Robots are gonna do
all the acting. Everything’s gonna be
cartoon AI stuff,” but that’s not what I feel like
we’re doing here at all. The point of this
is an exercise in thought. [Sharp] Okay,
stand by, everyone! Quiet, please! Action! [Goodwin]Making a machine
write like a person
is not about
replacing the person…
No, no, no! [Goodwin]
…it’s about augmenting
a person’s abilities.
[maniacal laughing] [Goodwin] It can empower people
to produce creative work that might be beyond
their native capacity. [screaming]
Come on! It’s wild to try to find
your interpretation
of this kind of text. Obviously, we usually
start with a script
that’s pretty coherent, and then I’ll break down
what the character says, and then I’ll decide, what are they feeling?
Why are they saying that? [Sharp]
Stay on her, stay on her. Just do that
walk-off again. Uh, stay where you are, John. Come back, Chelsey.
Do the walk-off again. This is harder for you,
and mo… and more frightening, and you’re checking
that he hasn’t gotten up. When AI
is writing the material, there isn’t any subtext. You realize what’s happening,
and you’re like, “Well, I’m gonna go
take refuge at the pillar. -Okay.
-All right? It stretches all of us.
It makes us all work harder. It’s one thing to bring
an existing script to life, and just do
your interpretation of it, but it’s another thing
to try to make it make sense, and then do
your interpretation of it. [Sharp] Let’s go
one more time. [Sharp] This Bobo character
that John is playing
is a fantasy figure, is this avatar of masculinity, is the sort of result
of watching
too many action films… but he’s confused, because he isn’t getting
the reaction that he expects. Action. [Chelsey screams] [whimpers in fear] [Sharp]
Somewhere in the script,
it talks about,
“Bobo leans over to Bobo.”
We think, “Oh, right, well,
let’s have a mirror…” No! You’re wrong! …and we can see
the two versions of Bobo,
for a moment, talking. [Sharp] Okay, let’s see you
in the mirror? [panting] Hey, did you get my money? Okay, great.
I’m getting terribly,
terribly happy. Some of that was so good. It was like such a go–
We’re, like, in a movie now. [Sharp]
Being surrounded with people who are throwing all of
their professional energy into something this ludicrous is just intrinsically enjoyable. They just breathe humanity into words that did not come
from a human being. -[man] All right,
let’s do it again.
-Okay, let’s do it. I think that making great art
requires human experience, but our human experience is now completely mapped
into data. This is where machine learning
keeps surprising us, is that it actually
has figured out stuff
that we didn’t realize it could. [Downey]Meaning, once
all our human experiences
are mapped into data,
AI will be able to mine it
for material
and make art?Look for patterns in
our happiness and heartbreak,
kick out a new song
or movie?
So this is all just this
one line of Benjamin writing, “putting on a show.” Right, right, right. [Sharp]So while
all that’s going on,
Girlfriend is on this couch,gradually waking up, right?-She’s in a horror movie…
-Right. -He’s in our action film.
-Oh! So in his head,
he’s having a wonderful
romantic time with her. Yeah, I love that. [Sharp] Do you remember his,
“Bobo leans over to Bobo”? -Mm-hmm.
-Remember that? So what we tried
to do for that is he looks in the mirror, and in the mirror,
it’s gonna be Osric. He’s created this avatar
version of himself, Bobo. -In a… in a…
-Okay, so that’s
the interpretation? -In that– Yeah, exactly.
-I like it. So this is what
these guys came up with. [panting] No! You’re wrong! You work really, really hard
to go, what’s a thing
that’s kinda coherent, that these actors can all
be performing one thing, we can all be making one thing, and we can say,
“This is what Benjamin meant?” -[Goodwin] Right.
-What does that
tell me about me? -[Goodwin] Right.
-Like, what… So… and what I already know
about me is I’m really antsy
about how much misogyny is kind of encoded into…
into culture. On one hand, you go,
“This is an important,
worthwhile thing to do–” On the other hand,
we’re projecting. And the other thing,
you’re projecting, -but we’re always projecting.
-Always. [Sharp] Literally,
all interpretation
is projection. [man] Take 6. [Goodwin] I like playing
with authorship, and people’s concepts
of authorship, and people’s concepts
of where fiction
and where ideas come from. [Sharp]
Generative screenwriting. Me and Ross started it. I don’t know if
it’s a new art form, but it’s a new chunk
of what cinema can be. That’s new. What should we do
next time, Ross? -Romantic comedy.
-Okay. [Downey]It’s hard to know
if machine learning
will ever decode
the mysteries of love
or creativity.
Maybe it’s not even a mystery,just data points,but what about
other human qualities,
like instinct?Driving a car
already requires us
to make countless
unconscious decisions.
AI is learning to do that,but can we teach it to do more?Racing is not just
driving a car.
It’s also about intuition,
caution, aggression,
and taking risks.[Steve]Holly,
can you confirm 200
at the end of this straight?
Okay. [Downey]It requires almost
a preternatural will to win.
So, how fast
can a racecar go…
without a human
behind the wheel?
[Holly Watson Nall]
Motorsport has always been taking technology
to the limits… [man] You all good
your side, Holly? Yeah, I’m ready to go. …and one of
the goals of Roborace
is to really facilitate the accelerated development
of driverless technology. Okay, so we’ll try
to launch again. [Watson Nall] By taking
the autonomous technology to the limits of its ability, we think that we can
develop the technology faster. [Downey]
British startup Roborace
wants to break new ground
in driverless cars.
To do so,
they believe they need
to test the boundaries
of the technology…
working at the very outer edge
of what’s safe and possible,
where the margin for error
is razor thin.
After years
of trial and error,
they’ve created
the world’s first AI racecar.
[Watson Nall] The thing
that I love most
about working at Roborace is we have a dream
of being faster, and better, and safer than a human. [Downey]More than
50 companies around the world
are working
to bring self-driving cars
to city streets.
The promise
of driverless taxis,
buses, and trucks
is transformative.It’ll make our world
safer and cleaner,
changing the way
our cities are designed,
societies function,even how we spend our time.[Martin Ford]Think about
a self-driving car
out in the real world.
In order to build that system
and have it work, it’s got to be
virtually perfect. If you had a 99% accuracy rate, that wouldn’t be
anywhere near enough, because once you take
that 1% error rate and you multiply that by
millions of cars on the road, I mean, you’d have accidents
happening constantly, so the error rate
has to be extraordinarily low in order to pull this off. [Downey]Roborace is betting
they can crack the code
by seeing just how far
the tech can go,
a place usually reserved for
only the best human drivers.
[Watson Nall]As a human,
you have lots of advantages
over a computer.
You know exactly
where you are in the world. You have eyes that can
enable you to see things, so we need to implement
technology on vehicles to enable them to see the world.We have a system called OxTS.It’s a differential GPS,which means
it’s military grade.
We also use LiDAR sensors.These are basically
laser scanners.
They create, for the vehicle,a 3D map of the world
around it.
And there’s one last thing
that we use,
between the cars.
Each of them
can tell the other car
the position of it
on the track.
And just to be clear,
your phone does not come
with military-grade GPS.
These cars?
Next level.
The challenging part
is to really fuse all
this information together. At Roborace,
we can provide the hardware, but then we need software
companies to come to us to implement their software. [Downey]Today,
Roborace has invited
two skilled teams
to test their latest
road rocket on the track.
[in English]
My name is Johannes. I’m from the Technical
University of Munich. I’m the project leader. Is the Wi-Fi working
off the car? I could check it. [Downey]T.U.M. from Germanyis one of the top technical
universities in Europe,
home to 17 Nobel Prize winners
in science.
I have no connection
to the car. Wifi doesn’t work. So we have no Wi-Fi
to the car… So we just need
to reset the router. My name is Max. I’m, uh… Uh, let’s figure out,
who am I? [chuckles] I’m, uh… In Arrival, I’m a product owner
of the self-driving system. [Downey]
Arrival is a UK startup
focused on designing
and building
next-gen electric vehicles
for commercial use.
Ah, okay, okay,
okay, good. [Downey]Each team created
their own custom software,
the AI driver
that pilots the car,
and since each of
the teams’ programmers
have their own distinct
does that mean
each of their AI drivers
will have different
personalities or instincts too?
[Danny King] The two teams
that we have here are using two slightly
different approaches to the same problem
of making a car
go ’round the track in the shortest distance
in the fastest way. [Watson Nall]
The T.U.M. strategy
is really to keep their code
as simple as possible.
It’s maybe a very German,
efficient way of doing things. Okay, thanks,
we will check now. [Watson Nall] Arrival’s code
is more complicated in that they use many more
of the sensors on the vehicle. It will be interesting to see whether it pays off
to be simple in your code, or slightly more complicated, to use more of
the functionality of the car. [Downey]The first test
for each team
is the overtake,to see if their AI
can pass another car
at high speed.[Betz]
It’s difficult for AI because we have to make
a lot of decisions, and a lot of planning,
a lot of computations to calculate
what the car should do
in which millisecond. [Watson Nall]Everybody
has seen high-speed crashes
in motorsport before.
We’d quite like to avoid that.For this reason,
during testing,
we keep a human in the car.[Steve speaking] [Reece speaking] Okay, Reece, enabling AI. Can you just confirm
you’ve got
the blue light, please? [Reece speaking] [Watson Nall]
In order to overtake, they need a second car on track
at the same time. This is a vehicle
that stays in human-driven mode
the whole time, so we know exactly
how it’s going to behave. Okay, launch AI
from the race control. And launching
in three, two, one. It’s really difficult for AI
to learn to overtake.When you have
one vehicle on track,
it only needs to make
decisions about itself,
but when you have two vehicles,you have the option
to create your behavior
in response to another vehicle.Okay, we are going
to release the speed limit
on your car now, Reece. [in English] Nice! Yeah, man. [Downey]Team T.U.M.has successfully completed
the overtake challenge.
Next up, team Arrival.So, Tim, can you go
to take position
on the start line, please? Enabling AICan you confirm
blue light, please?
[Tim speaking] And launch
in three, two, one. So it’s looking good so far. [Tim speaking] [tires screech] [thudding] Car crashed. Tim, can you
hear me? [Tim speaking] Has anyone got eyes
on what happened? [Tim speaking] Sorry, boys. [Bran Ferren]
Self-driving cars.
This is an idea that’s
been around since the ’30s,
hardly a new one.Why hasn’t it happened? It’s really hard.When there are unpredictable
things that happen,
that can get you
in a lot of trouble.
Now, sometimes trouble
just means it shuts down. Sometimes trouble means it gives you a result
that you weren’t expecting. I think he’s just… They’ve come back online
so aggressively… Plus or minus one G
coming back online. [Kumskoy] When the car
returned to the trajectory, it did it too aggressive, and actually steered
out of the racing track. -[King] My feeling
is that it overreacts.
-Yeah, yeah. So it’s not necessarily
the line that’s aggressive, it’s how it reacts
once it just gets
a little bit out of the line, and then overcorrects,
and then overcorrects. [Kumskoy] We were this close to really hitting the target
of our test, and it didn’t happen. It just slipped away,
so it was just… ah, disappointment. [King] There are so many
aspects of the car. The systems guys
have such a difficult job to make sure that everything
is absolutely perfect, because that’s what you need to be able to go
autonomous racing. Everything has to be perfect. [Downey]
Team Arrival’s program
just couldn’t hack it,
but for team T.U.M.,another test awaits…[Steve]Can we get the carinto the normal
start position, please?
[Downey]…and this next one
is all about speed.
Very high speed.The fastest that a human’s
ever driven around this track was 200 kph. [Downey]Translation,
that’s about 120 miles an hour.
[Watson Nall]So the AI
is gonna try to beat
that high speed.
And it’s gonna do it
without a human safety net,
because at that speed,it’s borderline unsafe
for people.
[Watson Nall]
When the driver climbs out
and shuts the door, yeah, your heart rate goes up. And we are launching in three, two, one. [accelerating]And launch successful.160. Next round, 200. [Downey]The car has six laps,six tries to hit top speed.Each lap,
the AI will increasingly
push the limits of control,
traction, and throttle,to break the human record.Holly, this is Steve.Can we confirm
in the atmos-data
it is safe to continue?[Watson Nall] Yeah, we think
it looks fairly controlled. [Steve]Okay, so the next run
should be V-max.
[Watson Nall]We have 210.That’s cool. [laughing] [Watson Nall] It was a real,
real sense of excitement to see it finally crack
the 210 kph mark. It was a real success
for Roborace functionality as well as building confidence
in the team’s software. It really showcases
what autonomous cars can do, not just on the racetrack, but also for everybody
around the world, so we’re really hoping that this will improve
road technology for the future. The current state of AI
is that there are some things that AI can really do
better than humans, and then there’s things that it can’t do
anywhere close to humans… but now where the frontier
is gonna be moving is where computers
come up to the human level, not quite,
and then surpass humans, and I think the odds
are overwhelming that we will eventually be able
to build an artificial brain that is at the level
of the human brain. The big question
is how long will it take? [Downey]“The hard problem.”It’s a philosophical phrasethat describes
difficult things to figure out,
like “the hard problem
of consciousness.”
We may never know
what consciousness is,
let alone if we can give it
to a machine,
but do we need to?What does a machine
really need to know
in order to be a good athlete,or an artist,or a lover?Will AI ever have
the will to win,
the depth to create,the empathy to connect
on a deep human level?
Maybe.Some say we’re just a bunch
of biological algorithms,
and that one day,evolution will evolve AI
to emulate humans
to be more like us…[bottle clatters]…or maybe it won’t…and human nature,
who we really are,
will remain a mystery.[servos whirring] We gave it some dialogue
to start with, like this line
fromSuperman.So you got some Superman/
Lois Lane stuff, huh? [Goodwin] Yeah,
so you wanna read it? [Sharp] Mm…
not that bit. Um, wait. Up, up, up, up, up.
Back up… okay. “Superman angrily grabs Lois
by the neck, slaps her against the wall,
and bares his teeth in fury.” “You’re wrong. You’re a grotesque
kind of monster.” -“You’re wrong!”
-“You’re a terrible liar.” “No! I’m sorry, I’m sorry. I can’t believe it!” “You’re so much more
than that, Lois.” “Please, please!” “How could you? No one can believe
who you are.” “Don’t be ridiculous. Please? How could you be
so much more than that?” “You’re such
a terrible liar. You can’t even believe
who you are. Please, unless you’re
really a no-good liar, you’re not even sure
if you’re good.” “Sorry, Superman,
I’m so sorry!” Superman is just not
making very much sense. Maybe kind of drunk
or something? In fact, it says,
“Superman isn’t funny. The two of them
are really different people. There is no such thing
as good good.” That’s pretty deep. “There is no such thing
as good good.” There is no such thing
as good good. So far as I know. Yeah.
Have you checked? I’m gonna Google it. -Can we Google it?


  1. Gabriel Barboza said:

    ta bom ne

    January 6, 2020
  2. UruguayoRandomTV said:

    Already written responses aren't ai

    January 6, 2020
  3. MiniLemmy said:

    Wait, what? Wrestler Jon Morrison (Jon Hennigan) is now starring in AI written movies??!!

    January 6, 2020
  4. Maiko Ye said:

    you know sex bots are coming

    January 6, 2020
  5. C-G-L Rockwell XI said:

    your sex robot is ready Dr. No

    January 6, 2020
  6. FGOKURULES said:

    HIS face looks like one of the dolls…. Or is it just me?

    January 6, 2020
  7. valia khalil said:

    OK did anyone see Osric Chau (aka Kevin Tran in supernatural) at 12:00 😊😍😘❤️

    January 6, 2020
  8. Black Parrot said:

    AI recommended this to me

    January 6, 2020
  9. Duggy Dugg said:

    i suspect a remote operator controlls Sophia..

    January 6, 2020
  10. XregularC said:

    I really struggle to see why the topic of robots being portrayed as female upsets people. Robots being portrayed as female does not objectify women but rather it "womanizes" objects (robots). I believe the reason we do this is that humans (or at least men) are biologically/instinctively wired to find female characteristics (such as appearance and voice) to be more calming, pleasant and nicer than that of male characteristics. If anything this highlights the fact that we men likes the presence of women and wants them around.

    January 6, 2020
  11. Sci Entia said:

    not intending to be disrespectful but i got the sense that the movie-guys didn't really understand how the AI works. it's literally clumping words together that stand in statistical relationship to one another according to the input texts. so all it does is glue-ing stereotypicial phrases together 😀
    i don't think that's creativity and i don't think there is any deeper meaning in this. but looking for meaning in any ever-so-non-sensical thing is an extremely human trait ; )

    January 6, 2020
  12. Frank Pon said:

    Lets restart the router 😅

    January 6, 2020
  13. Gandalf Spatz said:

    The only thing I would like to see more of in this show is giving credit to the organizations shown. Put their websites in the descriptions or even web pages where we can look more into their work. Links to what was shown per episode

    January 6, 2020
  14. The Destroyer of Leftists said:

    Detroit: Become human

    January 7, 2020
  15. Sivaral V said:

    Interesting. I wish to become AI scientist 🐎 from Tamilnadu

    To Create own AI military for me

    January 7, 2020
  16. Nick Prokopich said:

    (Bobo looks at bobo)

    January 7, 2020
  17. Amogh Kulkarni said:

    The problem is that AI models don't have the semantic knowledge of what they are creating (outputting if you will). And that's going to always make us humans (i.e. consumers of those artifacts) skeptical of the quality of AI-generated art.

    January 7, 2020
  18. dany roa said:

    Que tontería audio en ingle y subtítulos en inglés estoy pensando seriamente en dejar de pagar Premium no veo beneficios mejor nexflit

    January 7, 2020
  19. At Lanta said:

    Why don't you reveal how many times this show has been watched?

    January 7, 2020
  20. noddwyd said:

    The script writer has learned domestic violence, is what I got out of that part. That or the director loves cherry picking those parts out and using them.

    January 7, 2020
  21. D Man said:

    Chat bots are really ai ? Ai would be the chat bot just having words, and then learning to form its own sentences. Maybe ai needs a gibberish phase like babies before it could make sentences that make sense. An ai should review your input sentences and learn to speak, just like a baby. ai should not be solely a pre defined sentence selector thàt some programmer wrote, those sentences could just be initial training sentences, but the ai should ad its own training data based on new user sentences

    January 7, 2020
  22. DeLaPrd said:

    "Guys, here's what we could make but won't" – These kinds of documentary for at least 15 years.

    January 7, 2020
  23. Jayme Canaan said:

    Discordo, em se tratar de relacionamentos a I.A deve nos ajudar a manter o relacionamento com outras pessoas, e não substitui-los por maquinas.

    January 7, 2020
  24. Fabiano C said:

    When is the next episode going to be released?

    January 7, 2020
  25. stateofhibernation said:

    "There is no such thing as good good" you speak the true true

    January 7, 2020
  26. Lesi PlaysHacks said:

    this is free

    Yea. This is big brain time

    January 8, 2020
  27. Quix Nix said:

    I almost died when he said the thing about misogyny in culture… C E A S E

    January 8, 2020
  28. Rickey Bowers said:

    The principal problem with humans knowing where they are in the universe has always come down to ego – the consciousness is probably the same – we always like to imagine we're in control of a singular driving force. Whilst the reality is probably far from that.

    January 8, 2020
  29. moandjoaz said:

    I am really enjoying this series….can hardly wait for the next episodes

    January 8, 2020
  30. NihonNo Hoshi said:

    I'm just going to leave this here:

    This is a perverse abuse of technology, fueled by some very disturbed individuals who seem to not just want a sex doll, but one that can emulate a human. Very wrong, not only on a moral level but on an evolutionary one as well >_<

    January 9, 2020
  31. Lord k.Gaimiz said:

    7:12 the one with glasses reminded me of markus from Detroit become human

    January 9, 2020
  32. Monwabisi Mandla said:

    When are they realising the next episode?

    January 9, 2020
  33. Anonymous said:


    January 10, 2020
  34. Hirak Choudhury said:

    art + ai = artificial garbage

    January 10, 2020
  35. saravana Kumar said:

    Increase face emotion Like Sophia it become wonderful emotions expression robot

    January 10, 2020
  36. Xander 9898 said:

    Why did they remove likes and dislikes? This series is really good, and removing that looks bad.

    January 10, 2020
  37. тут ничего нет!! said:

    Вы круты!!!

    January 10, 2020
  38. alzathoth said:

    I would like to see what Benjamin would write for the last star wars film…

    January 11, 2020
  39. erina yuki said:

    they should interview Elon Musk about A.I…i know he disagree with A.I but…his team create A.I to beat Dota player. Loved to see him in this channel

    January 11, 2020
  40. Selva said:

    Who cares if a robot overtakes or not…what is the use to us ?? Create something helpful like the 2nd video of the series where it can cure eye from damaging permenantly that kind of stuff…not just making cyborg or robot to play as video game

    January 11, 2020
  41. Shayz said:

    Gatebox is way less creepy and is already a home assistant/companion that exists. No need to reinvent something

    January 11, 2020
  42. Damond Williams said:

    People watch races for 2 things. 1. Crashes. 2. Greatness. Humans love the danger, if you remove the danger then you remove the challenge hence you remove the race.

    January 11, 2020
  43. Haunted Thatcher said:

    I would like these guys to do a vid on the cyber body modders and the future of machine and purposefully removing limbs, eyes, organs to extend the human race.

    January 11, 2020
  44. Anxious Mac said:

    Writers all over the world lose their jobs.

    This industrial revolution 2, no job untouched.


    January 11, 2020
  45. icc2010 said:

    If anyone has seen the movie Bicentennial Man, with robin williams, this is basically that movie. Whn they showed the various heads of the AI, i heard robin say “i saw the inner me”. Very under rated movie

    January 11, 2020
  46. Tokent555 said:


    January 11, 2020
  47. Heather Faulkner said:

    When’s Shane Dawson gonna see this and flip??

    January 12, 2020
  48. Wiandry Adi Wasistio said:

    so this is all about detroit become human, astro boy, and J.A.R.V.I.S

    please don't go straight terminator

    January 12, 2020
  49. VP Vilakuang said:


    January 12, 2020
  50. Suhendar Aryadi said:

    Please Indonesia Subtitle

    January 12, 2020
  51. Rodrigo Cardoso said:

    Awesome is training your tensorflow model while watching this series!

    January 12, 2020
  52. Steve Legendzzx said:

    After watching ex machina and first half of this series, this gives me goosebumps 😬😬

    January 12, 2020
  53. AxillaMC said:

    Have these people seen Detroit Become Human?

    January 12, 2020
  54. twitch CYKA said:

    Почему русских субтитров нет?

    January 13, 2020
  55. Patrick Grenon said:

    it's John Hennigan! The mayor of SlamTown!! if you know you know.

    January 13, 2020
  56. Avinash Kachare said:

    Germans are amazing ! raise your hand if you agree!!

    January 13, 2020
  57. Harsh Barnwal said:

    Best series yet ♥️

    January 13, 2020
  58. Alkesh Sakre said:

    if any of the episode contain ai use in army then definately we are moving toward age of ultron and terminator .

    January 13, 2020
  59. ROSS Jones said:

    Can someone tell me when the next one is coming all YouTube originals is posting is that books thing but I want more of this

    January 13, 2020
  60. Luis Alejandro Fernandez Ruiz said:

    min 13:31 ´Art and creativity actually easier than problem-solving´ I do not agree with that, it is superb. The documentary is amazing, but first able if you want to be creative as Leonardo Davinci you need a complex mind.

    January 13, 2020
  61. YouTube Originals said:

    You spoke and we listened! Subtitles are now available in French, German, Hindi, Korean, Portuguese (BR) and Spanish for the first four episodes. Get ready for the final four episodes of Age of AI launching on January 15th!

    January 13, 2020
  62. JacobSk said:

    We all know where this eventually goes…

    Real world Anime Waifus and catgirls.

    Custom-design waifubot when?

    January 14, 2020
  63. Emik Zotic said:

    people try to make evi bot their gf dosent meant they actually mean it

    January 14, 2020
  64. Bedirhan Sasmaz said:

    Turn likes and dislikes on maynn

    January 14, 2020
  65. C H I R A G said:

    all the dolls are attractive not gonna lie.

    January 14, 2020
  66. Ryan Stark said:

    These machines will learn that humans are the cause of global destruction and will make a masterplan to eliminate humans.

    January 15, 2020
  67. Eros Camiseteria said:

    Fantastic video Youtube!!!!!!!!!!!!!!!!!!!!!!!!! Super!!!!!!!!!!!!

    January 15, 2020
  68. mirza arif said:

    Thoose dolls lil bit creepy..

    January 15, 2020
  69. Ashe Ent. said:


    January 15, 2020
  70. Mr Azza said:

    Million years from now, when our time is a lost or a forgetten era, and the machine will write a script where a new intelligent life will make this a religion

    January 15, 2020
  71. Randy St.fleur said:

    went from wwe to acting i see no career change

    January 16, 2020
  72. Samet Can Sönmez said:

    That's awesome.

    January 16, 2020
  73. Tech Addict said:

    As much as I love Deep Learning, I have to say that what they're showing here (Screenplay generation by AI) has 0% "creativity".
    Any one with knowledge of ML/AI will know this.
    Article generation or most other "AI" tasks are nothing but just copying what humans do, by taking an average of those records.

    GANs seem to have more like creativity, but the "AI"s shown in the video do not.

    January 16, 2020
  74. Uhu Die Eule said:

    They are cronos

    January 16, 2020
  75. mar montess said:


    January 16, 2020
  76. Galaxyer said:

    yay TUM

    January 16, 2020
  77. Subject OR36942 said:

    I've seen this it's not going to end well it leads to no emotions just fooling yourself just a monopoly like everything else

    January 16, 2020
  78. Paula Neilson said:

    Is it just me, or is the AI’s accent at 8.26 Scottish? Loving this series so far, it’s mind blowing how advanced and dedicated the creators are!

    January 16, 2020
  79. Hompanera Rocks said:

    Magway Is genius

    January 16, 2020
  80. Barry Allen said:

    A.I Become Human

    January 16, 2020
  81. Paul Bradley said:

    26,271,241 bytes = 26mb. That be . bro.

    January 16, 2020
  82. hyper YH said:

    ✌️cool series

    January 17, 2020
  83. Chosen One said:

    Me is something done with time

    January 17, 2020
  84. diego gamboa said:

    Thought i would never see the day tony stark trying to make a connection with a sex doll

    January 17, 2020
  85. Guilt Azaour said:

    you feed AI with Hollywood movies… why are you surprised ?? feed it books!

    January 17, 2020
  86. swampy said:

    I am so alone .. i need a bot but i need money for that .. forever alone

    January 17, 2020
  87. swampy said:

    Bobo made a boo boo. I am just kidding, just wanted to say it.

    January 17, 2020
  88. Sanja B said:

    First segment reminds me of movie "Her". Such a great invention.

    January 17, 2020
  89. Stan Petkov said:

    This actually makes me sad… We hate each other so we prefer fake perfected versions of us. Humans dont deserve humans….

    January 17, 2020
  90. Анна Никулина said:

    Зачем вам морды-лица на куске железа с пластиком?

    January 17, 2020
  91. Dele Mala said:

    so basically they're developing machine dolls who can "judge" in a way?

    January 17, 2020
  92. Martiddy - Sama said:

    14:32 That's a fancy way of saying 27 MegaBytes

    January 17, 2020
  93. David Betancourt Montellano said:

    "Debemos reiniciar el router" xD

    January 18, 2020
  94. Lucio Rivera said:

    normalizing the substitution of humans for machines?

    January 18, 2020
  95. Lucio Rivera said:

    normalizing the substitution of humans for machines?

    January 18, 2020
  96. joel dong said:

    can AI replace my wife. Can't wait for a "new model."

    January 18, 2020
  97. Pawan Bal said:

    Use cameras for eyes

    January 18, 2020
  98. Aniruddha R said:

    Why the doll looked like Jeff Bezos girlfriend ?

    January 18, 2020
  99. Aniruddha R said:

    Actress called A.I as "he"
    Liberals and LGBTQ + whatever people : how dare you

    January 18, 2020
  100. Shivam Parmar said:

    Why is the ai becoming so popular nowdays. Why now…? Please explain me anyone..

    January 18, 2020

Leave a Reply

Your email address will not be published. Required fields are marked *