Games & Quizzes
Don't forget to Sign In to save your points
This is a modal window.
Beginning of dialog window. Escape will cancel and close the window.
End of dialog window.
[MUSIC PLAYING]
Games & Quizzes
You may need to watch a part of the video to unlock quizzes
Don't forget to Sign In to save your points
PERFECT HITS | +NaN | |
HITS | +NaN | |
LONGEST STREAK | +NaN | |
TOTAL | + |
[MUSIC PLAYING]
[MUSIC PLAYING]
CHRIS KELLEY: Thank you so much for joining us.
CHRIS KELLEY: Thank you so much for joining us.
My name is Chris.
My name is Chris.
I'm a designer and prototyper working
I'm a designer and prototyper working
on immersive prototyping at Google,
on immersive prototyping at Google,
and I'm joined by Ellie and Luca.
and I'm joined by Ellie and Luca.
And today, we're going to talk about exploring AR interaction.
And today, we're going to talk about exploring AR interaction.
It's really awesome to be here.
It's really awesome to be here.
We explore immersive computing through rapid prototyping
We explore immersive computing through rapid prototyping
of AR and VR experiments.
of AR and VR experiments.
Often, that's focused on use case exploration or app ideas.
Often, that's focused on use case exploration or app ideas.
We work fast, which means we fail fast,
We work fast, which means we fail fast,
but that means that we learn fast.
but that means that we learn fast.
We spend a week or two on each prototyping sprint,
We spend a week or two on each prototyping sprint,
and at the end of the sprint, we end
and at the end of the sprint, we end
with a functional prototype starting
with a functional prototype starting
from a tightly scoped question.
from a tightly scoped question.
And then we put that prototype in people's hands
And then we put that prototype in people's hands
and we see what we can learn.
and we see what we can learn.
So this talk is going to be about takeaways we have
So this talk is going to be about takeaways we have
from those AR explorations.
from those AR explorations.
But first, I want to set the table a little bit
But first, I want to set the table a little bit
and talk about what we mean when we say augmented reality.
and talk about what we mean when we say augmented reality.
When a lot of people think about AR,
When a lot of people think about AR,
the first thing they think about is bringing virtual objects
the first thing they think about is bringing virtual objects
to users in the world.
to users in the world.
And it is that.
And it is that.
That's part of it.
That's part of it.
We call this the out of AR.
We call this the out of AR.
But AR also means more than that.
But AR also means more than that.
It means being able to understand the world visually
It means being able to understand the world visually
to bring information to users, and we call this understanding
to bring information to users, and we call this understanding
the in of AR.
the in of AR.
Many of the tools and techniques that
Many of the tools and techniques that
were created for computer vision and machine learning perfectly
were created for computer vision and machine learning perfectly
complement tools like ARCore, which is Google's AR
complement tools like ARCore, which is Google's AR
development platform.
development platform.
So when we explore AR, we build experiences
So when we explore AR, we build experiences
that include one of these approaches or both.
that include one of these approaches or both.
So this talk is going to be about three magic powers
So this talk is going to be about three magic powers
that we've found for AR.
that we've found for AR.
We think that these magic powers can help you build better AR
We think that these magic powers can help you build better AR
experiences for your users.
experiences for your users.
So we're going to talk about some prototypes that we've
So we're going to talk about some prototypes that we've
built and share our learnings with you
built and share our learnings with you
during each of these three magic power areas during the talk.
during each of these three magic power areas during the talk.
First, I'll talk to you about context-driven superpowers.
First, I'll talk to you about context-driven superpowers.
That's about how we can combine visual and physical
That's about how we can combine visual and physical
understanding of the world to make magical AR experiences.
understanding of the world to make magical AR experiences.
Then Ellie will talk to you about shared augmentations.
Then Ellie will talk to you about shared augmentations.
And this is really all about the different ways
And this is really all about the different ways
that we can connect people together in AR,
that we can connect people together in AR,
and how we can empower them just by putting them together.
and how we can empower them just by putting them together.
And then Luca will cover expressive inputs.
And then Luca will cover expressive inputs.
This is about how AR can help unlock
This is about how AR can help unlock
authentic and natural understanding for our users.
authentic and natural understanding for our users.
So let's start about context-driven superpowers.
So let's start about context-driven superpowers.
What this really means is using AR technologies
What this really means is using AR technologies
to deeply understand the context of a device,
to deeply understand the context of a device,
and then build experiences that directly leverage that context.
and then build experiences that directly leverage that context.
And there's two parts to an AR context.
And there's two parts to an AR context.
One is visual understanding, and the other
One is visual understanding, and the other
is physical understanding.
is physical understanding.
With ARCore, this gives your phone
With ARCore, this gives your phone
the ability to understand and sense
the ability to understand and sense
its environment physically.
its environment physically.
But through computer vision and machine learning,
But through computer vision and machine learning,
we can make sense of the world visually.
we can make sense of the world visually.
And by combining these results, we
And by combining these results, we
get an authentic understanding of the scene,
get an authentic understanding of the scene,
which is a natural building block of magical AR.
which is a natural building block of magical AR.
So let's start with visual understanding.
So let's start with visual understanding.
The prototyping community has done some awesome explorations
The prototyping community has done some awesome explorations
here, and we've done a few of our own
here, and we've done a few of our own
that we're excited to share.
that we're excited to share.
To start, we wondered if we could
To start, we wondered if we could
trigger custom experiences from visual signals in the world.
trigger custom experiences from visual signals in the world.
Traditional apps today leverage all kinds of device
Traditional apps today leverage all kinds of device
signals to trigger experiences.
signals to trigger experiences.
GPS, the IMU, et cetera.
GPS, the IMU, et cetera.
So could we use visual input as a signal as well?
So could we use visual input as a signal as well?
We built a really basic implementation of this concept.
We built a really basic implementation of this concept.
This uses ARCore and the Google Cloud Vision
This uses ARCore and the Google Cloud Vision
API that detects any kind of snowman
API that detects any kind of snowman
in the scene, which triggers a particle system that
in the scene, which triggers a particle system that
starts to snow.
starts to snow.
So through visual understanding, we
So through visual understanding, we
were able to tailor an experience to specific cues
were able to tailor an experience to specific cues
in the environment for users.
in the environment for users.
This enables adaptable and context aware applications.
This enables adaptable and context aware applications.
So even though this example is a simple one,
So even though this example is a simple one,
the concept can be extended so much further.
the concept can be extended so much further.
For example, yesterday we announced the augmented images
For example, yesterday we announced the augmented images
API for ARCore.
API for ARCore.
So if you use this, you can make something
So if you use this, you can make something
like an experience that reacts relative to device movement
like an experience that reacts relative to device movement
around an image in the scene, or even
around an image in the scene, or even
from a known distance to an object in the world.
from a known distance to an object in the world.
If you think this concept is interesting,
If you think this concept is interesting,
I highly recommend checking out the AR VR demo tent.
I highly recommend checking out the AR VR demo tent.
They have some amazing augmented images demos there.
They have some amazing augmented images demos there.
The next thing we wanted to know is
The next thing we wanted to know is
if we could bridge the gap between digital and physical,
if we could bridge the gap between digital and physical,
and, for example, bring some of the most delightful features
and, for example, bring some of the most delightful features
of e-readers to physical books.
of e-readers to physical books.
The digital age has brought all kinds of improvements
The digital age has brought all kinds of improvements
to some traditional human behaviors,
to some traditional human behaviors,
and e-readers have brought lots of cool new things to reading.
and e-readers have brought lots of cool new things to reading.
But if you're like me, sometimes you just
But if you're like me, sometimes you just
missed the tactility in holding a great book in your hands.
missed the tactility in holding a great book in your hands.
So we wanted to know if we could bridge that gap.
So we wanted to know if we could bridge that gap.
In this prototype, users highlight a passage or word
In this prototype, users highlight a passage or word
with their finger and they instantly
with their finger and they instantly
get back a definition.
get back a definition.
This is a great example of a short-form-focused interaction
This is a great example of a short-form-focused interaction
that required no setup for users.
that required no setup for users.
It was an easy win only made possible
It was an easy win only made possible
by visual understanding.
by visual understanding.
But as soon as we tried this prototype,
But as soon as we tried this prototype,
there were two downfalls that we noticed,
there were two downfalls that we noticed,
and they became immediately apparent when we used it.
and they became immediately apparent when we used it.
The first is that it was really difficult to aim your finger
The first is that it was really difficult to aim your finger
at a small moving target on a phone,
at a small moving target on a phone,
and maybe the page is moving as well,
and maybe the page is moving as well,
and you're trying to target this little word.
and you're trying to target this little word.
That was really hard.
That was really hard.
And the second was that when you're highlighting a word,
And the second was that when you're highlighting a word,
your finger is blocking the exact thing
your finger is blocking the exact thing
that you're trying to see.
that you're trying to see.
Now, these are easily solvable with a follow-up UX iteration,
Now, these are easily solvable with a follow-up UX iteration,
but they illustrate a larger lesson.
but they illustrate a larger lesson.
And that's that with any kind of immersive computing,
And that's that with any kind of immersive computing,
you really have to try it before you can judge it.
you really have to try it before you can judge it.
An interaction might sound great when you talk about it
An interaction might sound great when you talk about it
and it might even look good in a visual mock,
and it might even look good in a visual mock,
but until you have it in your hand
but until you have it in your hand
and you can feel it and try it, you're
and you can feel it and try it, you're
not going to know if it works or not.
not going to know if it works or not.
You really have to put it in a prototype
You really have to put it in a prototype
so you can create your own facts.
so you can create your own facts.
Another thing we think about a lot
Another thing we think about a lot
is, can we help people learn more effectively?
is, can we help people learn more effectively?
Could we use AR to make learning better?
Could we use AR to make learning better?
There's many styles of learning, and if you
There's many styles of learning, and if you
combine these styles of learning,
combine these styles of learning,
it often results in faster and higher-quality learning.
it often results in faster and higher-quality learning.
In this prototype, we combined visual, oral, verbal,
In this prototype, we combined visual, oral, verbal,
and kinesthetic learning to teach people how
and kinesthetic learning to teach people how
to make the perfect espresso.
to make the perfect espresso.
The videos explain--
The videos explain--
I'm sorry.
I'm sorry.
We placed videos around the espresso machine
We placed videos around the espresso machine
in the physical locations where that step occurs.
in the physical locations where that step occurs.
So if you were learning how to use the grinder,
So if you were learning how to use the grinder,
the video for the grinder is right next to it.
the video for the grinder is right next to it.
Now, for users to trigger that video,
Now, for users to trigger that video,
they move their phone to the area
they move their phone to the area
and then they can watch the lesson.
and then they can watch the lesson.
That added physical component of the physical proximity
That added physical component of the physical proximity
of the video and the actual device
of the video and the actual device
made a huge difference in general understanding.
made a huge difference in general understanding.
In our studies, users who had never used an espresso machine
In our studies, users who had never used an espresso machine
before easily made an espresso after using this prototype.
before easily made an espresso after using this prototype.
So for some kinds of learning, this
So for some kinds of learning, this
can be really beneficial for users.
can be really beneficial for users.
Now, unfortunately for our prototype,
Now, unfortunately for our prototype,
one thing that we learned here was
one thing that we learned here was
that it's actually really hard to hold your phone
that it's actually really hard to hold your phone
and make an espresso at the same time.
and make an espresso at the same time.
So you need to be really mindful of the fact
So you need to be really mindful of the fact
that your users might be splitting
that your users might be splitting
their physical resources between the phone and the world.
their physical resources between the phone and the world.
And so as it applies to your use case,
And so as it applies to your use case,
try building experiences that are really
try building experiences that are really
snackable and hands-free.
snackable and hands-free.
Speaking of combining learning and superpowers together,
Speaking of combining learning and superpowers together,
we wondered if AR could help us learn
we wondered if AR could help us learn
from hidden information that's layered in the world
from hidden information that's layered in the world
all around us.
all around us.
This is a prototype that we built
This is a prototype that we built
that's an immersive language learning app.
that's an immersive language learning app.
We showed translations roughly next to objects of interest
We showed translations roughly next to objects of interest
and positioned these labels by taking a point cloud
and positioned these labels by taking a point cloud
sample from around the object and putting the label sort
sample from around the object and putting the label sort
of in the middle of the points.
of in the middle of the points.
Users found this kind of immersive learning really fun,
Users found this kind of immersive learning really fun,
and we saw users freely exploring
and we saw users freely exploring
the world looking for other things to learn about.
the world looking for other things to learn about.
So we found that if you give people
So we found that if you give people
the freedom to roam and tools that are simple and flexible,
the freedom to roam and tools that are simple and flexible,
the experiences that you build for them
the experiences that you build for them
can create immense value.
can create immense value.
We now have physical understanding.
We now have physical understanding.
This is AR's ability to extract and infer
This is AR's ability to extract and infer
information and meaning from the world around you.
information and meaning from the world around you.
When a device knows exactly where it is, not only in space,
When a device knows exactly where it is, not only in space,
but also relative to other devices,
but also relative to other devices,
we can start to do things that really
we can start to do things that really
feel like you have superpowers.
feel like you have superpowers.
For example, we can start to make
For example, we can start to make
interactions that are extremely physical, natural,
interactions that are extremely physical, natural,
and delightful.
and delightful.
Humans have been physically interacting
Humans have been physically interacting
with each other for a really long time,
with each other for a really long time,
but digital life has abstracted some of those interactions.
but digital life has abstracted some of those interactions.
We wondered if we could swing the pendulum
We wondered if we could swing the pendulum
back the other direction a little bit using AR.
back the other direction a little bit using AR.
So in this prototype, much like a carnival milk bottle game,
So in this prototype, much like a carnival milk bottle game,
you fling a baseball out of the top of your phone
you fling a baseball out of the top of your phone
and it hits milk bottles that are shown on other devices.
and it hits milk bottles that are shown on other devices.
You just point the ball where you want to go, and it goes.
You just point the ball where you want to go, and it goes.
We did this by putting multiple devices
We did this by putting multiple devices
in a shared coordinate system, which
in a shared coordinate system, which
you could do using the new Google Cloud Anchors API
you could do using the new Google Cloud Anchors API
that we announced for ARCore yesterday.
that we announced for ARCore yesterday.
And one thing you'll notice here is
And one thing you'll notice here is
that we aren't even showing users past their camera.
that we aren't even showing users past their camera.
Now, we did that deliberately because we really
Now, we did that deliberately because we really
wanted to stretch and see how far we
wanted to stretch and see how far we
could take this concept of physical interaction.
could take this concept of physical interaction.
And one thing we learned was that once people learned
And one thing we learned was that once people learned
to do it, they found it really natural
to do it, they found it really natural
and actually had a lot of fun with it.
and actually had a lot of fun with it.
But almost every user that tried it had to be not only
But almost every user that tried it had to be not only
told how to do it, but shown how to do it.
told how to do it, but shown how to do it.
People actually had to flip this mental switch
People actually had to flip this mental switch
of the expectations they have for how a 2D smartphone
of the expectations they have for how a 2D smartphone
interaction works.
interaction works.
So you really need to be mindful of the context that people
So you really need to be mindful of the context that people
are bringing in and the mental models they have
are bringing in and the mental models they have
for 2D smartphone interactions.
for 2D smartphone interactions.
We also wanted to know if we could help someone visualize
We also wanted to know if we could help someone visualize
the future in a way that would let them make better decisions.
the future in a way that would let them make better decisions.
Humans pay attention to the things that matter to us.
Humans pay attention to the things that matter to us.
And in a literal sense, the imagery
And in a literal sense, the imagery
that appears in our peripheral vision
that appears in our peripheral vision
takes a lower cognitive priority than the things
takes a lower cognitive priority than the things
we're focused on.
we're focused on.
Would smartphone AR be any different?
Would smartphone AR be any different?
In this experiment, we overlaid the architectural mesh
In this experiment, we overlaid the architectural mesh
of the homeowner's remodel on top of the active construction
of the homeowner's remodel on top of the active construction
project.
project.
The homeowner could visualize in context
The homeowner could visualize in context
what the changes to their home was going to look like.
what the changes to their home was going to look like.
Now, at the time that this prototype was created,
Now, at the time that this prototype was created,
we had to do actual manual alignment of this model
we had to do actual manual alignment of this model
on top of the house.
on top of the house.
You could do it today.
You could do it today.
If I rebuilt it, I would use the augmented images API
If I rebuilt it, I would use the augmented images API
that we announced yesterday.
that we announced yesterday.
It would be much easier to put a fixed
It would be much easier to put a fixed
image in a location, the house, and sync them together.
image in a location, the house, and sync them together.
But even with that initial friction for the UX,
But even with that initial friction for the UX,
the homeowner got tremendous value out of this.
the homeowner got tremendous value out of this.
In fact, they went back to their architect after seeing this
In fact, they went back to their architect after seeing this
and changed the design of their new home
and changed the design of their new home
because they found out that they weren't going
because they found out that they weren't going
to have enough space in the upstairs bathroom-- something
to have enough space in the upstairs bathroom-- something
they hadn't noticed in the plans before.
they hadn't noticed in the plans before.
So the lesson is that if you provide people high-quality,
So the lesson is that if you provide people high-quality,
personally relevant content, you can create ways
personally relevant content, you can create ways
that people will find really valuable and attention grabbing
that people will find really valuable and attention grabbing
experiences.
experiences.
But when does modifying the real environment
But when does modifying the real environment
start to break down?
start to break down?
You may be familiar with the uncanny valley.
You may be familiar with the uncanny valley.
It's a concept that suggests when
It's a concept that suggests when
things that are really familiar to humans
things that are really familiar to humans
are almost right but just a little bit off,
are almost right but just a little bit off,
it makes us feel uneasy.
it makes us feel uneasy.
Subtle manipulations of the real environment in AR
Subtle manipulations of the real environment in AR
can sometimes feel similar.
can sometimes feel similar.
It can be difficult to get right.
It can be difficult to get right.
In this specific example, we tried
In this specific example, we tried
removing things from the world.
removing things from the world.
We created this AR invisibility cloak for the plant.
We created this AR invisibility cloak for the plant.
What we did was we created a point cloud around the object,
What we did was we created a point cloud around the object,
attached little cubes to the point cloud,
attached little cubes to the point cloud,
applied a material to those points,
applied a material to those points,
and extracted the texture from the surrounding environment.
and extracted the texture from the surrounding environment.
That worked pretty well in uniform environments,
That worked pretty well in uniform environments,
but unfortunately, the world doesn't have too many of those.
but unfortunately, the world doesn't have too many of those.
It's made up of dynamic lighting and subtle patterns,
It's made up of dynamic lighting and subtle patterns,
so this always ended up looking a little bit weird.
so this always ended up looking a little bit weird.
Remember to be thoughtful about the way that you add
Remember to be thoughtful about the way that you add
or remove things from the environment.
or remove things from the environment.
People are really perceptive, and so you
People are really perceptive, and so you
need to strive to build experiences
need to strive to build experiences
that align with their expectations,
that align with their expectations,
or at the very least, don't defy them.
or at the very least, don't defy them.
But is physical understanding always critical?
But is physical understanding always critical?
All points in the section have their place,
All points in the section have their place,
but, ultimately, you have to be guided by your critical user
but, ultimately, you have to be guided by your critical user
journeys.
journeys.
In this example, we wanted to build
In this example, we wanted to build
a viewer for this amazing 3D model by Damon [INAUDIBLE]..
a viewer for this amazing 3D model by Damon [INAUDIBLE]..
It was important that people could see the model in 3D
It was important that people could see the model in 3D
and move around to discover the object.
and move around to discover the object.
A challenge, though, was that the camera feed
A challenge, though, was that the camera feed
was creating a lot of visual noise and distraction.
was creating a lot of visual noise and distraction.
People were having a hard time appreciating
People were having a hard time appreciating
the nuances of the model.
the nuances of the model.
We adopted concepts from filmmaking and guided users
We adopted concepts from filmmaking and guided users
by using focus and depth of field,
by using focus and depth of field,
all which were controlled by the user's motion.
all which were controlled by the user's motion.
This resulted in people feeling encouraged to explore,
This resulted in people feeling encouraged to explore,
and they really stopped getting distracted
and they really stopped getting distracted
by the physical environment.
by the physical environment.
So humans are already great at so many things.
So humans are already great at so many things.
AR really allows us to leverage those existing capabilities
AR really allows us to leverage those existing capabilities
to make interactions feel invisible.
to make interactions feel invisible.
If we leverage visual and physical understanding
If we leverage visual and physical understanding
together, we can build experiences that
together, we can build experiences that
really give people superpowers.
really give people superpowers.
With that, Ellie is going to talk to you
With that, Ellie is going to talk to you
about special opportunities we have in shared augmentations.
about special opportunities we have in shared augmentations.
ELLIE NATTINGER: Thanks, Chris.
ELLIE NATTINGER: Thanks, Chris.
So I'm Ellie Nattinger.
So I'm Ellie Nattinger.
I'm a software engineer and prototyper
I'm a software engineer and prototyper
on Google's VR and AR team.
on Google's VR and AR team.
Chris has talked about the kinds of experiences
Chris has talked about the kinds of experiences
you start to have when your devices can understand
you start to have when your devices can understand
the world around you, and I'm going
the world around you, and I'm going
to talk about what happens when you can share those experiences
to talk about what happens when you can share those experiences
with the people around you.
with the people around you.
We're interested not only in adding AR augmentations
We're interested not only in adding AR augmentations
to your own reality, but also in sharing those augmentations.
to your own reality, but also in sharing those augmentations.
If you listened to the developer keynote yesterday,
If you listened to the developer keynote yesterday,
you know that shared AR experiences
you know that shared AR experiences
is a really big topic for us these days.
is a really big topic for us these days.
For one thing, a shared reality lets
For one thing, a shared reality lets
people be immersed in the same experience.
people be immersed in the same experience.
Think about a movie theater.
Think about a movie theater.
Why do movie theaters exist?
Why do movie theaters exist?
Everybody's watching a movie that they could probably
Everybody's watching a movie that they could probably
watch at home on their television or their computer
watch at home on their television or their computer
by themselves much more comfortably not
by themselves much more comfortably not
having to go anywhere, but it feels qualitatively
having to go anywhere, but it feels qualitatively
different to be in a space with other people sharing
different to be in a space with other people sharing
that experience.
that experience.
And beyond those kinds of shared passive experiences,
And beyond those kinds of shared passive experiences,
having a shared reality lets you collaborate, lets you learn,
having a shared reality lets you collaborate, lets you learn,
lets you build and play together.
lets you build and play together.
We think you should be able to share your augmented realities
We think you should be able to share your augmented realities
with your friends, and your families, and your colleagues,
with your friends, and your families, and your colleagues,
so we've done a variety of explorations
so we've done a variety of explorations
about how do you build those kinds of shared
about how do you build those kinds of shared
realities in AR.
realities in AR.
First, there's kind of a technical question.
First, there's kind of a technical question.
How do you get people aligned in a shared AR space?
How do you get people aligned in a shared AR space?
There's a number of ways we've tried.
There's a number of ways we've tried.
If you don't need a lot of accuracy,
If you don't need a lot of accuracy,
you could just start your apps with all the devices
you could just start your apps with all the devices
in approximately the same location.
in approximately the same location.
You could use markers or augmented images
You could use markers or augmented images
so multiple users can all point their devices at one picture
so multiple users can all point their devices at one picture
and get a common point of reference--
and get a common point of reference--
cures the zero, zero, zero of my virtual world.
cures the zero, zero, zero of my virtual world.
And you can even use the new ARCore Cloud Anchors API
And you can even use the new ARCore Cloud Anchors API
that we just announced yesterday to localize
that we just announced yesterday to localize
multiple devices against the visual features
multiple devices against the visual features
of a particular space.
of a particular space.
In addition to the technical considerations,
In addition to the technical considerations,
we've found three axes of experience
we've found three axes of experience
that we think are really useful to consider
that we think are really useful to consider
when you're designing these kinds of shared
when you're designing these kinds of shared
augmented experiences.
augmented experiences.
First of those is co-located versus remote.
First of those is co-located versus remote.
Are your users in the same physical space
Are your users in the same physical space
or different physical spaces?
or different physical spaces?
Second is, how much precision is required, or is it optional?
Second is, how much precision is required, or is it optional?
Do you have to have everybody see the virtual bunny
Do you have to have everybody see the virtual bunny
at exactly the same point in the world,
at exactly the same point in the world,
or do you have a little bit of flexibility about that?
or do you have a little bit of flexibility about that?
And the third is whether your experience
And the third is whether your experience
is synchronous or asynchronous.
is synchronous or asynchronous.
Is everybody participating in this augmented experience
Is everybody participating in this augmented experience
at exactly the same time, or at slightly different times?
at exactly the same time, or at slightly different times?
And we see these not as necessarily binary axes,
And we see these not as necessarily binary axes,
but more of a continuum that you can
but more of a continuum that you can
consider when you're designing these multi-person AR
consider when you're designing these multi-person AR
experiences.
experiences.
So let's talk about some prototypes and apps that
So let's talk about some prototypes and apps that
fall on different points of the spectrum and the lessons
fall on different points of the spectrum and the lessons
we've learned from them.
we've learned from them.
To start with, we've found that when
To start with, we've found that when
you've got a group that's interacting
you've got a group that's interacting
with the same content in the same space,
with the same content in the same space,
you really need shared, precise, spatial registration.
you really need shared, precise, spatial registration.
For example, let's say you're in a classroom.
For example, let's say you're in a classroom.
Imagine if a group of students who
Imagine if a group of students who
are doing a unit on the solar system could all look at
are doing a unit on the solar system could all look at
and walk around the globe, or an asteroid field,
and walk around the globe, or an asteroid field,
or look at the sun.
or look at the sun.
In Expeditions AR, one of Google's initial AR
In Expeditions AR, one of Google's initial AR
experiences, all the students can point their devices
experiences, all the students can point their devices
to a marker, they calibrate themselves
to a marker, they calibrate themselves
against a shared location, they see the object
against a shared location, they see the object
in the same place, and then what this allows
in the same place, and then what this allows
is for a teacher to be able to point out
is for a teacher to be able to point out
particular parts of the object.
particular parts of the object.
Oh, if you all come over and look at this side of the sun,
Oh, if you all come over and look at this side of the sun,
you see a cut-out into its core.
you see a cut-out into its core.
Over here on the Earth, you can see a hurricane.
Over here on the Earth, you can see a hurricane.
Everybody starts get a spatial understanding
Everybody starts get a spatial understanding
of the parts of the object and where they are in the world.
of the parts of the object and where they are in the world.
So when does it matter that your shared
So when does it matter that your shared
space has a lot of precision?
space has a lot of precision?
When you have multiple people who
When you have multiple people who
are all in the same physical space
are all in the same physical space
interacting with or looking at the exact same
interacting with or looking at the exact same
augmented objects at the same time.
augmented objects at the same time.
We were also curious--
We were also curious--
how much can we take advantage of people's existing spatial
how much can we take advantage of people's existing spatial
awareness when you're working in high-precision shared spaces?
awareness when you're working in high-precision shared spaces?
We experimented with this in this multi-person construction
We experimented with this in this multi-person construction
application, where you've got multiple people who
application, where you've got multiple people who
are all building onto a shared AR object in the same space.
are all building onto a shared AR object in the same space.
Adding blocks to each other, everybody's
Adding blocks to each other, everybody's
being able to coordinate.
being able to coordinate.
And you want to be able to tell what part of the object someone
And you want to be able to tell what part of the object someone
is working on.
is working on.
Have your physical movement support that collaboration.
Have your physical movement support that collaboration.
Like, if Chris is over here and he's
Like, if Chris is over here and he's
placing some green blocks in the real world,
placing some green blocks in the real world,
I'm not going to step in front of him
I'm not going to step in front of him
and start putting yellow blocks there instead.
and start putting yellow blocks there instead.
We've got a natural sense of how to collaborate, how to arrange,
We've got a natural sense of how to collaborate, how to arrange,
how to coordinate ourselves in space.
how to coordinate ourselves in space.
People already have that sense.
People already have that sense.
So we can keep that in a shared AR
So we can keep that in a shared AR
if we've got our virtual objects precisely lined up enough.
if we've got our virtual objects precisely lined up enough.
We also found it helpful to notice
We also found it helpful to notice
that because you can see both the digital object but also
that because you can see both the digital object but also
the other people through the pass-through camera,
the other people through the pass-through camera,
you are able to get a pretty good sense of what people
you are able to get a pretty good sense of what people
were looking at as well as what they were interacting with.
were looking at as well as what they were interacting with.
We've also wondered what would it
We've also wondered what would it
feel like to have a shared AR experience for multiple people
feel like to have a shared AR experience for multiple people
in the same space, but who aren't necessarily interacting
in the same space, but who aren't necessarily interacting
with the same things?
with the same things?
So think of this more like an AR LAN party.
So think of this more like an AR LAN party.
Where we're all in the same space,
Where we're all in the same space,
or maybe could be different spaces,
or maybe could be different spaces,
we're seeing connected things, and we're
we're seeing connected things, and we're
having a shared experience.
having a shared experience.
So this prototype's a competitive quiz guessing game
So this prototype's a competitive quiz guessing game
where you look at the map and you
where you look at the map and you
have to figure out where on the globe you think is represented
have to figure out where on the globe you think is represented
and stick your pushpin in, get points
and stick your pushpin in, get points
depending on how close you are.
depending on how close you are.
We've got the state synced, so we know who's winning.
We've got the state synced, so we know who's winning.
But the location of where that globe is
But the location of where that globe is
doesn't actually need to be synchronized.
doesn't actually need to be synchronized.
And maybe you don't want it to be synchronized because I don't
And maybe you don't want it to be synchronized because I don't
want anybody to get a clue based on where I'm sticking
want anybody to get a clue based on where I'm sticking
my pushpin into the globe.
my pushpin into the globe.
It's fun to be together, even when we're not looking
It's fun to be together, even when we're not looking
at exactly the same AR things.
at exactly the same AR things.
And do we always need our spaces to align exactly?
And do we always need our spaces to align exactly?
Sometimes it's enough just to be in the same room.
Sometimes it's enough just to be in the same room.
This prototype example's of an AR boat race.
This prototype example's of an AR boat race.
You blow on the microphone of your phone,
You blow on the microphone of your phone,
and it creates the wind that propels your boat
and it creates the wind that propels your boat
down the little AR track.
down the little AR track.
By us being next to each other when we start the app
By us being next to each other when we start the app
and spawn the track, we get a shared physical experience
and spawn the track, we get a shared physical experience
even though our AR worlds might not perfectly align.
even though our AR worlds might not perfectly align.
We get to keep all the elements of the social game play--
We get to keep all the elements of the social game play--
talking to each other, our physical presence--
talking to each other, our physical presence--
but we're not necessarily touching the same objects.
but we're not necessarily touching the same objects.
Another super interesting area we've been playing with
Another super interesting area we've been playing with
is how audio can be a way to include
is how audio can be a way to include
multiple people in a single device AR experience.
multiple people in a single device AR experience.
If you think of the standard Magic Window device AR,
If you think of the standard Magic Window device AR,
it's a pretty personal experience.
it's a pretty personal experience.
I'm looking at this thing through my phone.
I'm looking at this thing through my phone.
But now, imagine you can leave a sound in AR that
But now, imagine you can leave a sound in AR that
has a 3D position like any other virtual thing,
has a 3D position like any other virtual thing,
and now you start to be able to hear it,
and now you start to be able to hear it,
even if you're not necessarily looking at it.
even if you're not necessarily looking at it.
And other people can hear the sound
And other people can hear the sound
from your device at the same time.
from your device at the same time.
So for an example, let's say you could leave little notes
So for an example, let's say you could leave little notes
all over your space.
all over your space.
Might look something like this.
Might look something like this.
I'm a plant.
I'm a plant.
I'm a plant.
I'm a plant.
I'm a plant.
I'm a plant.
I'm elephant.
I'm elephant.
I'm elephant.
I'm elephant.
I'm elephant.
I'm elephant.
This is a chair.
This is a chair.
This is a chair.
This is a chair.
This is a chair.
This is a chair.
I'm a plant.
I'm a plant.
I'm a plant.
I'm a plant.
I'm elephant.
I'm elephant.
I'm elephant.
I'm elephant.
This is a chair.
This is a chair.
This is a chair.
This is a chair.
So notice, you don't have to be the one with a phone
So notice, you don't have to be the one with a phone
to get a sense of where these audio annotations start
to get a sense of where these audio annotations start
to live in physical space.
to live in physical space.
Another question we've asked--
Another question we've asked--
if you have a synchronous AR experience with multiple people
if you have a synchronous AR experience with multiple people
who are in different places, what kind of representation
who are in different places, what kind of representation
do you need of the other person?
do you need of the other person?
So let's imagine you have maybe a shared AR photos app
So let's imagine you have maybe a shared AR photos app
where multiple people can look at photos
where multiple people can look at photos
that are arranged in space.
that are arranged in space.
So I'm taking pictures in one location,
So I'm taking pictures in one location,
I'm viewing them arranged around me in AR,
I'm viewing them arranged around me in AR,
and then I want to share my AR experience
and then I want to share my AR experience
with Luca, who comes in and joins me
with Luca, who comes in and joins me
from a remote location.
from a remote location.
What we found-- we needed a couple of things to make
What we found-- we needed a couple of things to make
us feel like we were connected and sharing the same AR
us feel like we were connected and sharing the same AR
experience, even though we were in different places.
experience, even though we were in different places.
We needed to have a voice connection so we could actually
We needed to have a voice connection so we could actually
talk about the pictures, and we needed
talk about the pictures, and we needed
to know where the other person was looking.
to know where the other person was looking.
See which picture you're paying attention to when
See which picture you're paying attention to when
you're talking about it.
you're talking about it.
But what was interesting is we didn't actually
But what was interesting is we didn't actually
need to know where the other person was, as long as we had
need to know where the other person was, as long as we had
that shared frame of reference.
that shared frame of reference.
We're all here, here's what I'm looking at,
We're all here, here's what I'm looking at,
here's what Luca's looking at.
here's what Luca's looking at.
We've also been curious about asymmetric experiences.
We've also been curious about asymmetric experiences.
What happens when users share the same space
What happens when users share the same space
and the same augmentations, but they've got different roles
and the same augmentations, but they've got different roles
in the experience?
in the experience?
So for instance, in this prototype,
So for instance, in this prototype,
Chris is using his phone as a controller to draw in space,
Chris is using his phone as a controller to draw in space,
but he's not actually seeing the AR annotations he's drawing.
but he's not actually seeing the AR annotations he's drawing.
The other person sees the same AR content
The other person sees the same AR content
and uses their phone to take a video.
and uses their phone to take a video.
They're playing different roles in the same experience.
They're playing different roles in the same experience.
Kind of artist versus cinematographer.
Kind of artist versus cinematographer.
And we found there could be some challenges
And we found there could be some challenges
to asymmetric experiences if there's
to asymmetric experiences if there's
a lack of information about what the other person is
a lack of information about what the other person is
experiencing.
experiencing.
For instance, Chris can't tell what Luca's filming
For instance, Chris can't tell what Luca's filming
or see how his drawing looks from far away.
or see how his drawing looks from far away.
So as we mentioned previously, these kinds
So as we mentioned previously, these kinds
of different combinations of space,
of different combinations of space,
and time, and precision are relevant for multi-person AR
and time, and precision are relevant for multi-person AR
experiences, and they have different technical and
experiences, and they have different technical and
experiential needs.
experiential needs.
If you have multiple people in the same space
If you have multiple people in the same space
with the same augmentations at the same time,
with the same augmentations at the same time,
then you need a way of sharing.
then you need a way of sharing.
You need a way of common localization.
You need a way of common localization.
That's why we created the new Cloud Anchors API.
That's why we created the new Cloud Anchors API.
If you've got multiple people in the same space
If you've got multiple people in the same space
with different augmentations at the same time,
with different augmentations at the same time,
the kind of AR LAN party model, you
the kind of AR LAN party model, you
need some way to share data.
need some way to share data.
And if you've got multiple people
And if you've got multiple people
in different spaces interacting with the same augmentations
in different spaces interacting with the same augmentations
at the same time, you need sharing
at the same time, you need sharing
in some kind of representation of that interaction.
in some kind of representation of that interaction.
So shared AR experiences is a big area.
So shared AR experiences is a big area.
We've explored some parts of the space.
We've explored some parts of the space.
We'd love to see what you all come up with.
We'd love to see what you all come up with.
So Chris has talked about examples
So Chris has talked about examples
where your device understands your surroundings
where your device understands your surroundings
and gives you special powers, I talked about examples
and gives you special powers, I talked about examples
where you've got multiple people who
where you've got multiple people who
can collaborate and interact.
can collaborate and interact.
Now Luca will talk about what happens when your devices have
Now Luca will talk about what happens when your devices have
a better understanding of you and allow
a better understanding of you and allow
for more expressive inputs.
for more expressive inputs.
Luca?
Luca?
LUCA PRASSO: Thank you, Ellie.
LUCA PRASSO: Thank you, Ellie.
My name is Luca Prasso, and I'm a prototyper
My name is Luca Prasso, and I'm a prototyper
and a technical artist working in the Google AR and VR team.
and a technical artist working in the Google AR and VR team.
So let's talk about the device that you
So let's talk about the device that you
carry with you every day and the ones that are all around you,
carry with you every day and the ones that are all around you,
and how they can provide the meaningful and authentic
and how they can provide the meaningful and authentic
signals that we can use in our augmented experiences.
signals that we can use in our augmented experiences.
So ARCore tracks the device motion
So ARCore tracks the device motion
as we move to the real world and provides some understanding
as we move to the real world and provides some understanding
of the environment.
of the environment.
And these signals can be used to create powerful, and creative,
And these signals can be used to create powerful, and creative,
and expressive tools, and offer new ways for us
and expressive tools, and offer new ways for us
to interact with digital content.
to interact with digital content.
So the data represents who we are, what we know,
So the data represents who we are, what we know,
and what we have.
and what we have.
And we were interested in understanding
And we were interested in understanding
if the user can connect more deeply if the data is displayed
if the user can connect more deeply if the data is displayed
around them in 3D, and through AR and physical aspirations,
around them in 3D, and through AR and physical aspirations,
they can look at this data.
they can look at this data.
So we took a database of several thousand world cities,
So we took a database of several thousand world cities,
and we mapped it in an area that's
and we mapped it in an area that's
wide as a football field.
wide as a football field.
We assign a dot to every city and we scale the dot based
We assign a dot to every city and we scale the dot based
on the population of the city.
on the population of the city.
And each country has a different color.
And each country has a different color.
So now you can walk to this data field.
So now you can walk to this data field.
And as ARCore tracks the motion of the user,
And as ARCore tracks the motion of the user,
we play footsteps in sync.
we play footsteps in sync.
You take a step and you hear a step.
You take a step and you hear a step.
And [INAUDIBLE] sound fields surrounds the user
And [INAUDIBLE] sound fields surrounds the user
and enhances the experience and the sense of exploration
and enhances the experience and the sense of exploration
of this data forest.
of this data forest.
And flight paths are displayed up in the sky.
And flight paths are displayed up in the sky.
And the pass-through camera is heavily tinted
And the pass-through camera is heavily tinted
so that we can allow the user to focus on the data
so that we can allow the user to focus on the data
and then still give a sense of presence.
and then still give a sense of presence.
And what happens is the user, as he walks to the physical space,
And what happens is the user, as he walks to the physical space,
he starts mapping, and pairing, and creating
he starts mapping, and pairing, and creating
this mental map between the data and the physical location.
this mental map between the data and the physical location.
And starts understanding better, in this particular case,
And starts understanding better, in this particular case,
the relative distance between the places.
the relative distance between the places.
And what we discover is also that the gestures that
And what we discover is also that the gestures that
are a part of our digital life every day, a pinch to zoom,
are a part of our digital life every day, a pinch to zoom,
it's now in AR something more traditional.
it's now in AR something more traditional.
It's actually moving closer to the digital object
It's actually moving closer to the digital object
and inspecting it like we do with a real object.
and inspecting it like we do with a real object.
And pan and drag means taking a couple of steps
And pan and drag means taking a couple of steps
to the right to look at the information.
to the right to look at the information.
So physical exploration like this is very fascinating,
So physical exploration like this is very fascinating,
but we need to take into account all the different users
but we need to take into account all the different users
and provide the alternative move and affordances.
and provide the alternative move and affordances.
So in AR, a user can move everywhere,
So in AR, a user can move everywhere,
but what if he cannot or he doesn't want to move?
but what if he cannot or he doesn't want to move?
What if he's sitting?
What if he's sitting?
So in this particular case, we allow the user to simply point
So in this particular case, we allow the user to simply point
the phone everywhere they want to go, tap on the screen
the phone everywhere they want to go, tap on the screen
anywhere, and the application will move the point of view
anywhere, and the application will move the point of view
in that direction.
in that direction.
At the same time, we still have to provide audio, haptics,
At the same time, we still have to provide audio, haptics,
and color effects to enhance the sense of physical space
and color effects to enhance the sense of physical space
the user has to have while traveling.
the user has to have while traveling.
And so we found that this is a powerful mechanism
And so we found that this is a powerful mechanism
to explore a certain type of data that makes sense in the 3D
to explore a certain type of data that makes sense in the 3D
space and to allow the user to discover hidden patterns.
space and to allow the user to discover hidden patterns.
But can we go beyond the pixels that you
But can we go beyond the pixels that you
can find on your screen?
can find on your screen?
We're fascinated by the spatial audio and a way
We're fascinated by the spatial audio and a way
to incorporate audio into an AR experience.
to incorporate audio into an AR experience.
So we combine ARCore and the Google Resonance SDK.
So we combine ARCore and the Google Resonance SDK.
And Resonance is this very powerful spatial audio engine
And Resonance is this very powerful spatial audio engine
that recently Google open-sourced.
that recently Google open-sourced.
And you should check it out because it's great.
And you should check it out because it's great.
And so now I can take audio sources
And so now I can take audio sources
and place them into the 3D locations,
and place them into the 3D locations,
and animate them, and describe the properties of the walls,
and animate them, and describe the properties of the walls,
and the ceilings, and the floor, and all the obstacles.
and the ceilings, and the floor, and all the obstacles.
And now as the ARCore moves the point of view,
And now as the ARCore moves the point of view,
it carries with it the digital ears,
it carries with it the digital ears,
the Resonance used to render accurately
the Resonance used to render accurately
the sounds in the scene.
the sounds in the scene.
So what can we do with this?
So what can we do with this?
So we imagine, what if I can sit next to a performer
So we imagine, what if I can sit next to a performer
during an acoustic concert, or a classical concert, or a jazz
during an acoustic concert, or a classical concert, or a jazz
performance?
performance?
What if I can be onstage with actors,
What if I can be onstage with actors,
and listen to their play, and be there?
and listen to their play, and be there?
So we took two amazing actors, Chris and Ellie,
So we took two amazing actors, Chris and Ellie,
and we asked them to record separately
and we asked them to record separately
lines from Shakespeare.
lines from Shakespeare.
And we placed these audio sources a few feet apart
And we placed these audio sources a few feet apart
and we surrounded the environment
and we surrounded the environment
with an ambisonic sound field of a rain forest, of the raining.
with an ambisonic sound field of a rain forest, of the raining.
And then later on, we switched to a room with a lot of reverb
And then later on, we switched to a room with a lot of reverb
into the walls.
into the walls.
CHRIS KELLEY: Thou told'st me they were stolen unto this
CHRIS KELLEY: Thou told'st me they were stolen unto this
wood, and here I am, and wode within this wood,
wood, and here I am, and wode within this wood,
because I cannot meet my Hermia.
because I cannot meet my Hermia.
Hence, get thee gone, and follow me no more.
Hence, get thee gone, and follow me no more.
ELLIE NATTINGER: You draw me, you hard-hearted adamant,
ELLIE NATTINGER: You draw me, you hard-hearted adamant,
but yet you draw not iron, for my heart is true as steel.
but yet you draw not iron, for my heart is true as steel.
Leave you your power to draw, and I shall
Leave you your power to draw, and I shall
have no power to follow you.
have no power to follow you.
CHRIS KELLEY: Do I entice you?
CHRIS KELLEY: Do I entice you?
Do I speak to you fair?
Do I speak to you fair?
Or rather, do I not in plainest truth tell you, I do not,
Or rather, do I not in plainest truth tell you, I do not,
nor I cannot love you?
nor I cannot love you?
LUCA PRASSO: So now the user can walk around,
LUCA PRASSO: So now the user can walk around,
maybe with his eyes closed, a nice pair of headphones,
maybe with his eyes closed, a nice pair of headphones,
and it's like being on stage with these actors.
and it's like being on stage with these actors.
So we took this example and we extended it.
So we took this example and we extended it.
We observed that we can build in real-time a 2D map of where
We observed that we can build in real-time a 2D map of where
the user has been so far with his phone
the user has been so far with his phone
as he's walking around.
as he's walking around.
And so at any given time when the user hits a button,
And so at any given time when the user hits a button,
we can programmatically place audio recording in space
we can programmatically place audio recording in space
where we know that the user can reach with the phone
where we know that the user can reach with the phone
and with their ears.
and with their ears.
[MUSIC PLAYING]
[MUSIC PLAYING]
And suddenly, the user becomes the human mixer
And suddenly, the user becomes the human mixer
of this experience.
of this experience.
And different instruments can populate
And different instruments can populate
your squares, and your rooms, and your schools.
your squares, and your rooms, and your schools.
And this opens the door to an amazing amount
And this opens the door to an amazing amount
of opportunities with AR audio-first experiments.
of opportunities with AR audio-first experiments.
So let's go back to visual understanding.
So let's go back to visual understanding.
Chris mentioned that the computer vision and machine
Chris mentioned that the computer vision and machine
learning can interpret the things that are around us,
learning can interpret the things that are around us,
and this is also important to understand the body in turning
and this is also important to understand the body in turning
into an expressive controller.
into an expressive controller.
So in real life, we are surrounded
So in real life, we are surrounded
by a lot of sound sources for all of the places.
by a lot of sound sources for all of the places.
And naturally, our body and our head
And naturally, our body and our head
moves to mix and focus on what we like
moves to mix and focus on what we like
and what we want to listen to.
and what we want to listen to.
So can we take this intuition into the way we watch movies
So can we take this intuition into the way we watch movies
or play video games on a mobile device?
or play video games on a mobile device?
So what we did, we took the phone camera signal,
So what we did, we took the phone camera signal,
fed it to Google Mobile Vision.
fed it to Google Mobile Vision.
That gave us a head position and head orientation.
That gave us a head position and head orientation.
And we fed it to Google Resonance SDK.
And we fed it to Google Resonance SDK.
And we said, OK, you're watching a scene in which actors
And we said, OK, you're watching a scene in which actors
are in a forest, and they're all around you, and it's raining.
are in a forest, and they're all around you, and it's raining.
So now as I leave my phone far away from my head,
So now as I leave my phone far away from my head,
I hear the forest.
I hear the forest.
As I'm taking the phone closer to my face,
As I'm taking the phone closer to my face,
I start hearing the actors playing.
I start hearing the actors playing.
I warn you, this is an Oscar performance.
I warn you, this is an Oscar performance.
[THUNDER RUMBLES]
[THUNDER RUMBLES]
ELLIE NATTINGER: Our company here.
ELLIE NATTINGER: Our company here.
CHRIS KELLEY: My man, according to the script.
CHRIS KELLEY: My man, according to the script.
ELLIE NATTINGER: Here is the scroll of every man's name
ELLIE NATTINGER: Here is the scroll of every man's name
which is thought fit through all Athens to play in our interlude
which is thought fit through all Athens to play in our interlude
before the duke and the duchess on his [INAUDIBLE]
before the duke and the duchess on his [INAUDIBLE]
LUCA PRASSO: So now what is interesting is
LUCA PRASSO: So now what is interesting is
that the tiny little motions that we
that the tiny little motions that we
can do when we're watching and we're playing this experience,
can do when we're watching and we're playing this experience,
it can be turned into subtle changes in the user
it can be turned into subtle changes in the user
experience that we can control.
experience that we can control.
So we talk about how the changes in poses
So we talk about how the changes in poses
can become a trigger to drive interaction.
can become a trigger to drive interaction.
In this Google Research app called [INAUDIBLE],,
In this Google Research app called [INAUDIBLE],,
we actually exploit the opposite--
we actually exploit the opposite--
the absence of motion.
the absence of motion.
And when the user--
And when the user--
in this case, my kids--
in this case, my kids--
stop posing, the app takes a picture.
stop posing, the app takes a picture.
And so the simple mechanism that is triggered by computer vision
And so the simple mechanism that is triggered by computer vision
creates the incredible, delightful opportunities
creates the incredible, delightful opportunities
that, apparently, my kids love.
that, apparently, my kids love.
And Research is doing incredible progress
And Research is doing incredible progress
in looking at an RGB image and understanding where
in looking at an RGB image and understanding where
the body pose and skeleton is.
the body pose and skeleton is.
And you should check out the Google Research blog post
And you should check out the Google Research blog post
because their post estimation research is amazing.
because their post estimation research is amazing.
So we took Ellie's video and we fed it to the machine computer
So we took Ellie's video and we fed it to the machine computer
algorithm.
algorithm.
And we got back, a bunch of 3D poses and segmentation
And we got back, a bunch of 3D poses and segmentation
masks of Ellie.
masks of Ellie.
And this opens the door to a lot of variety
And this opens the door to a lot of variety
of experiments with creative filters
of experiments with creative filters
that we can apply to this.
that we can apply to this.
But what's more interesting for us
But what's more interesting for us
is that it also allows us to understand better
is that it also allows us to understand better
the intent and the context of the user.
the intent and the context of the user.
So we took this pose estimation technology
So we took this pose estimation technology
and we added a digital character.
and we added a digital character.
Now it tries to mimic what the human character is doing.
Now it tries to mimic what the human character is doing.
And this allows [INAUDIBLE] now to bring your family
And this allows [INAUDIBLE] now to bring your family
and friends--
and friends--
in this case, my son, Noah--
in this case, my son, Noah--
into the scene so that he can act and create a nice video.
into the scene so that he can act and create a nice video.
But this also, like Ellie mentioned before,
But this also, like Ellie mentioned before,
we should consider the situation,
we should consider the situation,
because this is an asymmetric experience.
because this is an asymmetric experience.
What you don't see here is how frustrated
What you don't see here is how frustrated
my son was after a few minutes because he
my son was after a few minutes because he
couldn't see what was going on.
couldn't see what was going on.
I was the one having fun taking picture and video him,
I was the one having fun taking picture and video him,
and he didn't see much.
and he didn't see much.
He could only hear the lion roaring.
He could only hear the lion roaring.
So we need to be extremely mindful
So we need to be extremely mindful
as the developer about this unbalance of delight.
as the developer about this unbalance of delight.
And so maybe I should have passed the image of the phone
And so maybe I should have passed the image of the phone
to a nearby TV so I can make my son first-class citizen
to a nearby TV so I can make my son first-class citizen
in this experience.
in this experience.
So all this AR technology and the physical
So all this AR technology and the physical
and the visual understanding are ingredients
and the visual understanding are ingredients
that allow us to unlock all kinds of new expressive input
that allow us to unlock all kinds of new expressive input
mechanisms.
mechanisms.
And we are still exploring.
And we are still exploring.
We're just at the beginning of this journey.
We're just at the beginning of this journey.
But we are excited to hear what you think
But we are excited to hear what you think
and what you want to come up with.
and what you want to come up with.
So to summarize, we shared a bunch
So to summarize, we shared a bunch
of ways in which we think about AR and various aspirations
of ways in which we think about AR and various aspirations
that we have done.
that we have done.
We talked about expanding our definition of AR.
We talked about expanding our definition of AR.
Putting content into the world, but also pulling information
Putting content into the world, but also pulling information
from the world.
from the world.
And these are all ingredients that we
And these are all ingredients that we
use to create these magical AR superpowers to enhance
use to create these magical AR superpowers to enhance
the social interactions and to express yourself
the social interactions and to express yourself
in this new digital medium.
in this new digital medium.
So we combined ARCore capabilities
So we combined ARCore capabilities
with different Google technologies,
with different Google technologies,
and this gives us the opportunity
and this gives us the opportunity
to explore all these new interaction models.
to explore all these new interaction models.
And we encourage you, developers,
And we encourage you, developers,
to stretch your definition of AR.
to stretch your definition of AR.
But we want to do this together.
But we want to do this together.
We're going to keep exploring, but what
We're going to keep exploring, but what
we want to hear what tickled you,
we want to hear what tickled you,
what tickled your curiosity.
what tickled your curiosity.
So we can wait to see what you build next.
So we can wait to see what you build next.
Thank you very much for coming.
Thank you very much for coming.
[MUSIC PLAYING]
[MUSIC PLAYING]
[MUSIC PLAYING]. CHRIS KELLEY: Thank you so much for joining us.. My name is Chris.. I'm a designer and prototyper working. on immersive prototyping at Google,. and I'm joined by Ellie and Luca.. And today, we're going to talk about exploring AR interaction.
It's really awesome to be here.. We explore immersive computing through rapid prototyping
of AR and VR experiments.. Often, that's focused on use case exploration or app ideas.
We work fast, which means we fail fast,. but that means that we learn fast.. We spend a week or two on each prototyping sprint,. and at the end of the sprint, we end. with a functional prototype starting. from a tightly scoped question.. And then we put that prototype in people's hands. and we see what we can learn.. So this talk is going to be about takeaways we have
/THro͞o/
continuing or valid to final destination. From the beginning of something until the end. Allowing you to pass between, or to.
/dəˈzīnər/
Describing something that has been designed. person who plans look or workings of something prior to it being made.
/ˌôɡˈmentəd/
having been made greater in size or value. To make something larger, stronger, more effective.
/iˌmərsiv/
(of computer display or system) generating three-dimensional image which appears to surround user.
/kəmˈpyo͞odiNG/
use or operation of computers. To find out by calculating or estimating.
/ˈfōkəs/
With your attention, concentration on one thing. To see clearly by adjusting your eyes or a camera.
Metric | Count | EXP & Bonus |
---|---|---|
PERFECT HITS | 20 | 300 |
HITS | 20 | 300 |
STREAK | 20 | 300 |
TOTAL | 800 |
Sign in to unlock these awesome features: