Placeholder Image

字幕表 動画を再生する

  • Cool.

  • Hello, everyone.

  • Hey.

  • Good morning.

  • Thanks for coming here.

  • It's probably been a long week.

  • The last couple of days have been crazy here,

  • and the traffic's not been fun either.

  • So hopefully it's been working all well for you guys.

  • Some of you have made your way from a different parts

  • of the country, many different areas.

  • But I hope it's been worth it.

  • And there's a lot of interesting things that we've been doing.

  • So we want to talk a little bit about some of them.

  • You've heard our keynotes and the great products we have,

  • et cetera.

  • Today I'm going to talk a little bit about machine learning,

  • right?

  • The title says machine learning is not

  • the future, which is kind of weird

  • given that we're talking about machine learning in everything

  • we're doing here.

  • But the point that I wanted to bring across here

  • is it's not really the future.

  • It's part of everything we have today

  • as you're seeing without products as all the things

  • that we're talking about.

  • We're using machine learning today

  • in everything we're doing.

  • And I want to talk about some of that

  • and tell you how you can do that too.

  • But first, let's start with this all started.

  • So and this goes back to 1997.

  • I don't know how many of you remember this Tik-Tok

  • from "The Wizard of Oz."

  • He's one of the early versions of robots in modern times,

  • really.

  • Although there are references to automatons going back

  • to the 5th Century B.C. And stuff,

  • even the Egyptians and others way back when.

  • But in the modern times, this is probably

  • one of the earliest references to something

  • that was close to a robot.

  • Any of you familiar with Tik-Tok?

  • How many of you are familiar with Tik-Tok?

  • Not too many.

  • Let me tell you a little bit about him.

  • So this one is-- in fact, the term robot

  • wasn't even coined back when Tik-Tok

  • was created by Frank Baum.

  • And what this robot would do is it had a winding thing.

  • And you needed to wind the robot.

  • And it would run.

  • And it could do pretty much what a human could do but not

  • really be alive.

  • And it was a great reference to things

  • that we've always wanted to do.

  • And so a lot of this AI and robotics

  • and a lot of these things have been part of science fiction

  • for a very, very long time.

  • In fact, a lot of what we do today in science

  • has been driven by people, literally authors, et cetera,

  • are able to really build up and think about what

  • the future might be like.

  • This is one example, but there are many, many more.

  • For example, if some of you have read Asimov,

  • he has a bunch of books.

  • And he has this whole robotics cities,

  • again way back in the '30s where he talks about robots

  • and how they might be like and the kind

  • of things that come with it.

  • And that was very interesting.

  • A lot of people over the years have

  • been inspired by these science fiction books and movies

  • to do a lot of interesting things.

  • Another example of that is back in the '70s

  • from "The Hitchhikers Guide to the Galaxy"

  • there is this robot called Marvin,

  • which is this depressed-- it's always depressing.

  • He's always talking about these things he's not happy about.

  • He's just too smart for all the things happening it.

  • And it's really a great example of what

  • people have been thinking about what's

  • going to happen in the future.

  • And now going to another example,

  • "Star Trek," another thing.

  • The example that I put here, you can

  • go to Data in "Star Trek" who talks about-- this

  • is a poem he composes, which I thought was pretty funny,

  • and how people have been thinking about AIs

  • and what they would do and what they

  • would be doing in the future.

  • In this case, again, it's like a robot,

  • like a humanoid, somebody who can do a lot of things,

  • but is not quite human.

  • Then coming to more recent times, more closer to 2001,

  • if you've seen Steven Spielberg's movie "AI" called

  • "AI" itself, they have these mechanical robots

  • that can really do-- that look like humans, that

  • can do all kinds of things.

  • But they don't feel.

  • And then they build this little kid David,

  • who can actually feel and love as well,

  • and really changes how the perception is

  • and how they think about AI or what it means

  • to the people around them.

  • And then much more recently there's

  • this movie called "Her" in 2013, which

  • talks about AI without the shape, without the robotics,

  • and all of that.

  • But AI is just that live within your computer as an OS,

  • essentially, that can interact with you.

  • That's your assistant, but much more than that.

  • It actually feels as well.

  • It understands things, et cetera.

  • So all these are great.

  • There are so many different things

  • that people have been talking about,

  • people have been thinking about.

  • But these are still science fiction.

  • This is not really what machine learning is about today.

  • This has always been the future.

  • It's still the future.

  • Maybe at some point it will be a reality.

  • But that's not where we are today.

  • That's not what we're talking about.

  • But there are some real things that we can do today.

  • And those are some of the things that I'm

  • going to talk about today.

  • So we've actually made over the last few years

  • real progress in terms of all the different things that we

  • can do with AI with machine learning,

  • use all the products that you see around you.

  • And I'm going to talk a little bit about those.

  • There's so much in there that has

  • benefited from machine learning from what we call AI as well.

  • And so this is some of the smallest of products, really,

  • at Google that we made that used machine learning in some ways.

  • But at Google, anytime we think of a product

  • there's, of course, programming.

  • And you build it, and you do all sorts of things with it.

  • But machine learning is an integral part of everything

  • we do in building that, because we want these products to be

  • really smart to give you the right things,

  • to not just follow your actions, but really

  • give you the right things when you want them as you want them.

  • And I'll go over some examples of these in later slides

  • as well.

  • So before I go into other things,

  • let's just go a little bit into what deep learning

  • and machine learning is about.

  • And I'm just quickly going to give you some examples

  • on a website that we have.

  • So part of this slide is going to talk about TensorFlow.

  • And recently we put the site up called

  • playground.tensorflow.org that allows

  • you to play with neural nets that allows you to really do

  • different kinds of things, and allows you to really understand

  • how these networks work, how machine learning works,

  • and be able to play with some of those problems.

  • So I'm going to start with a very, very basic problem

  • classification.

  • The goal in this case is there are two kinds of clients.

  • I just want to classify that it's ARB, in this case

  • the orange or the blue ones.

  • And I'm going to use the very simplest case.

  • It's a very simple linear classification,

  • if any of you guys know what that is.

  • But the idea is you have some inputs.

  • In this case, the x-axis and the y-axis are inputs.

  • And based on those two inputs you

  • want to decide if it's a blue one or an orange one.

  • And so what this model is going to learn

  • is some parameters to figure out-- OK.

  • If I get some point, I get the an x1 value and a x2 value,

  • how do I decide that?

  • It's some mathematical computation

  • that it needs to do.

  • So in this case, it does that iteratively.

  • Let's just quickly run this.

  • I've just set it up as a linear model.

  • So lets just run it very quick.

  • And as you see here, its looking at x1 and x2.

  • And very, very quickly it figures out

  • how to separate those out.

  • It just draw a line in the middle.

  • And its just optimizing what that line looks

  • like, how to separate these out very simply.

  • So this was a very easy problem.

  • This is typically not how the real problems are,

  • but it's a great place to start.

  • So let's go to a slightly more complex problem now.

  • So let's go to this one here where

  • you still have those two kinds of points,

  • but now they're clustered differently.

  • So there's one in the center-- is a circle,

  • and the rest are around it.

  • So can we actually solve this using the same kind of methods?

  • Can we actually classify using the same thing?

  • If I run this, do you think it would work?

  • Any guesses?

  • Let's try it out.

  • So it's trying to separate this out.

  • It's basically trying to draw a line.

  • But it's lost.

  • It's really not making a progress

  • if you see what's happening here.

  • So there are a few things you can do about this.

  • One, in terms of classification, again,

  • going a little bit into machine learning into the details,

  • you can add more features, more kinds of things about the input

  • data that helped the model understand

  • or how to separate them.

  • So if you guys remember your math back from high school

  • or college, in this particular case,

  • given how these clusters are structured,

  • if you looked at a couple more features

  • from based on the inputs, like in this case,

  • x1 squared and x2 squared, you can probably

  • use those to separate this out better.

  • So let's try this.

  • Let's add these two and see what happens.

  • And taking those two values, now it

  • knows how to separate those and clearly

  • separate it out into two separate, different classes.

  • So this works.

  • This great great, right?

  • This is what happens or used to happen

  • in machine learning very often for every single thing

  • that you do.

  • If you take a bunch of inputs, you want to solve some problem.

  • It's typically the inputs themselves

  • that are not in the same kind of mode

  • that you need to solve the problem.

  • And what you do is called feature generation

  • where you are combining or crossing

  • those features in different ways to really solve your problem.

  • But there's a lot of work you need

  • to do in figuring out the right features that make sense.

  • In this case, it was a simple problem.

  • So we knew x squared, x1 squared or x2 squared would work.

  • But in some cases, the problem is not as simple

  • and you really need to do a little bit more.

  • So for those kind of cases, we have

  • something called deep learning which

  • seems to work really well.

  • So let's go to go back to this example

  • and let's see if we can solve it in a different way as well.

  • So let's take out these two features first.

  • Let's change this linear to a nonlinearity.

  • In this case, I'm going to pick correctify linear unit.

  • And I'm going to add a hidden layer with a few more neurons.

  • So what this is doing is it has some inputs, as we

  • said in this case just x1 x2.

  • I'm adding another layer, which is basically

  • trying to combine those inputs in whatever

  • way is interesting to solve our particular problem.

  • And then let's see if we can train this to actually solve

  • the same problem as well.

  • So if you see this, we took the same features, no feature

  • generation.

  • We just added one layer.

  • And this is a simple problem, so one layer works.

  • And it's very easily able to separate those out.

  • This is essentially what deep learning is doing for you.

  • As you add more layers, it can understand and learn

  • the right layers of abstraction that

  • are the most interesting to solve your current problem.

  • And of course, it depends on the problem.

  • It doesn't have to be a simple classification.

  • It could be regression.

  • It could be recommendation, whatever you are trying to do,

  • really.

  • You can apply similar techniques to do that.

  • So let's go back to our slides.

  • In a sense, deep learning's basically

  • the same machine learning ideas with more layers

  • there that allow you to get different kinds of abstraction.

  • Of course, as part of that you get many, many more parameters.

  • You learn things differently.

  • This example is actually for what

  • we call inception retreat, which is our state of the art image

  • model for image classification.

  • So And we'll go through some examples of how we use it.

  • But this basically gives you something

  • nearer human accuracy.

  • So machine learning has been around for a long time,

  • in fact way more than 10 years.

  • What's really changed since then?

  • One of them is more compute, definitely.

  • Over the years, our computers have gotten faster.

  • We have more of them.

  • We've been using newer chips, CPUs, GPUs, now 2PUs here.

  • And so a lot of things that we're doing here--

  • and more computer allows us to really build

  • more complex models, more bigger models.

  • But that needs a lot of other things

  • to go with it as well for that to work, in this case,

  • better algorithms and more data help as well.

  • So over the years we've improved the algorithms slowly.

  • A lot of these algorithms that we talk about in deep learning

  • are 30, 40, 50 years old.

  • But over the last 10 years, we've

  • seen some improvements in the right places.

  • For example, in a case of deep learning,

  • there are some improvements to back propagation

  • or the kind of nonlinearities that we use that really make it

  • much easier to optimize those.

  • It used to be really, really hard

  • to train some of these models, even if you had the compute.

  • Even if you had the data.

  • Now it's much easier and there are better techniques

  • to do that.

  • So the math and the algorithms is improved.

  • And of course, the amount of data that we have,

  • that keeps growing.

  • And that really is something that machine learning can help.

  • If you think about it with small amounts of data humans

  • are great.

  • You can maybe read 10 pages of a book

  • and understand that and sift it down.

  • But what if you had a million pages?

  • You can't really expect a human to go through those million

  • pages, understand those, and summarize it for you.

  • You really need some machines that can automatically

  • do this for you.

  • And that's really where machine learning comes in.

  • So now let's talk a little bit about what's

  • happening in research in machine learning now.

  • We've talked a little bit about science-fiction,

  • mentioned a bunch of products that are using this.

  • And I'll go a little bit more into the products later.

  • But let's talk a little bit about the research, where this

  • is going, some of the examples of where

  • things of made improvements in the recent past.

  • So the first one I want to talk about is AlphaGo.

  • How many of you of heard of AlphaGo?

  • Wow, quite a few.

  • So if you remember actually almost 20 years ago now,

  • there was deep blue, which is a similar chess computer.

  • In some sense it was similar that the goal was, OK.

  • Can we really beat the best humans,

  • Casper of the reigning champion in that time using computers?

  • And back in 1997 was the first time

  • Deep Blue, which is this computer built by IBM

  • that actually beat the world champion,

  • beat Kasparov in a full championship match.

  • That was pretty amazing.

  • And that was 20 years ago.

  • It took a long time to go from there to Go,

  • partly because Go itself as a game is

  • way more complex than chess is.

  • The possibilities that you have at every level at every step

  • are much, much bigger, much more than what chess is.

  • And so one of the differences there

  • is with Deep Blue you could get away

  • with some smarts and a lot of brute force.

  • You could have a large supercomputer.

  • You have some basic smarts and a lot of rules around-- there

  • are a lot of chess players there that

  • actually taught Deep Blue in the sense it would train them

  • on-- OK.

  • Here the good moves.

  • Here are not, et cetera, et cetera,

  • plus allowing it to do a lot of brute force

  • in trying out-- OK, looking ahead how many different

  • moves you can do.

  • Now those same techniques just could not

  • work for Go because of the different possibilities

  • at every step.

  • At every step we can do a really large number of moves,

  • and supporting that or being able to go past a few steps

  • is extremely hard, even for modern day computers.

  • So what really worked in this case

  • was a combination of smartness and the computer.

  • And in case of smartness, they applied

  • deep learning to figure out, OK, looking at a board, what kind

  • of moves make the most sense?

  • And then let's try them out and see what makes sense

  • and which ones of these play well.

  • And so it's a combination of improved algorithms, improved

  • machine learning, combined with the compute power as well.

  • Another example is ImageNet, which

  • is this data set of a million images with 1,000

  • classes of different kinds.

  • In fact, a lot of classes are so close

  • together that it's very hard even

  • for humans who don't know about them to identify

  • without learning.

  • For example, there are breeds of dogs which look

  • very similar, at least to me.

  • For people who really know dogs, of course, it's easier.

  • Over the last few years, over the last five years,

  • especially on this data set, this

  • as really driven the state of, well, vision understanding

  • for machines is.

  • So about five years ago, the accuracy

  • that computers would get was maybe 70%, 75%.

  • This was the measured top five accuracy, which is basically

  • is the most-- the correct answer among the top five results

  • that the computer predicts?

  • Now, five years ago this was something like 75%.

  • Today it's past.

  • It's over 96%, which is considered to be better

  • than human accuracy.

  • Humans also make mistakes in this kind of thing.

  • So it's really coming a long way and just shows

  • the kind of progress we can make if we really get down

  • to improving machine learning.

  • This one's an interesting one, taking the kind of image models

  • that we've trained and really letting them dream.

  • This is what we call Deep Dream.

  • There are some folks at Google who

  • basically took those new networks

  • and wanted to understand what's happening in these networks.

  • What are they learning right?

  • And so in this case, what they did

  • was took a few neurons in the network,

  • just let them train from that point

  • and try to see what we can generate

  • from those neurons, what they might be understanding

  • from the images.

  • And so they converged to some of these.

  • So in some of these images in the top right,

  • it's basically started from some kind of palace or whatever

  • and combined those, and [INAUDIBLE] of that.

  • The one in the top left it has some fountains in there

  • and a bunch of other things that sort of combined.

  • And this is very interesting because in some ways

  • there are similarities to what people do as well.

  • Of course, this is a very different view.

  • But it's starting to learn in many, many ways.

  • Machine learning is really doing the same kind of learning

  • that people might do.

  • So now I'm going to give you some examples.

  • Let me just go to this and play some, a couple more

  • videos for you.

  • In this case, it's doing what-- it's basically taking the video

  • and applying different kinds of painting styles.

  • In this particular one it's just picking a blue painting style

  • and applying it to the existing video in real time.

  • Now it's a charcoal sketch.

  • This is just an example of how we're

  • mixing machine learning with things like art

  • to all kinds of interesting things in

  • and interesting combinations things that you

  • can do with these.

  • Here's another one where it makes it look like a newspaper.

  • So this one's based on a paper based on

  • by Justin Johnson and others here at Sanford.

  • Let me give you another example of similar things

  • that we can do.

  • So this one is interesting where the network

  • learned from a bunch of Chinese characters.

  • It looked at how they're drawn, and was

  • able to really draw completely new Chinese characters I've

  • never seen before.

  • These aren't real characters, actually,

  • for those of you who understand the script.

  • I don't, but then they look very real.

  • Even for people who understand the script or who

  • know the script, they feel like they're almost there,

  • they're clearly not the right character or anything

  • that they've ever seen.

  • But it's very understanding what's happening there.

  • It gets tagged it's supposed to look like this

  • and is able to understand that whole domain as well.

  • I'm going to show you a couple more videos.

  • Here's one where it's basically learned from a lot of numbers.

  • These are house numbers from Street View.

  • Let's just go back here and play it and pause it at the end.

  • And so it's learned a bunch of-- it's looked at house numbers

  • from Street View and is able to really generate those.

  • It's actually never seen these particular numbers.

  • The network is generating them.

  • This is the kind of things that these networks can do today.

  • And the last one, in this case, it's actually generating faces.

  • And so, again, it learned from a large face data set

  • and now is able to generate faces in all kinds of ways.

  • Again, these are not faces that's seen directly.

  • It's regenerating them based on all it's understanding

  • of what faces look like.

  • Yeah.

  • Another one that we've actually released very recently

  • is this model to parse sentences.

  • So this is a very semantic parsing of sentences

  • where it looks at a sentence.

  • In this case, I booked a ticket to Google,

  • and tries to understand what the subjects are,

  • all the grammatical stuff about that,

  • actually things that I personally don't understand,

  • and it's hard for me to learn.

  • But this is able to actually learn

  • from a lot of sentences that were given to it.

  • In this example, it's actually making the sentence piece

  • by piece, and then building a tree of hierarchy,

  • and doing the parsing in this case.

  • This kind of thing is extremely useful

  • when you're trying to understand natural language.

  • And this just goes to show how much these computers can

  • understand the language that you and I speak in some cases

  • better than humans can as well or at least better than kids

  • can.

  • Here's another example, sort of going back to the data example

  • where in science fiction this-- here's this sort of poem

  • that a computer might generate.

  • This is actually something that was generated by a real program

  • today or a real model today.

  • It's clearly not great, but in this particular case,

  • it just learned from a bunch of books.

  • It read those books or understood those books.

  • And was able to generate this particular poem.

  • So now let's talk about this grateful research.

  • Now let's talk about a few products

  • and how we're using them.

  • So we had this slide early on where

  • I showed you a bunch of products that [INAUDIBLE] using this.

  • Let's go into some of those products.

  • Let's see how we are actually using them.

  • What are the kind of things we're doing with machine

  • learning in these products?

  • First a search-- I'm sure most of you

  • have-- or probably all of you have used search

  • or are using search in some ways.

  • This is a very good example of how machine learning is used.

  • The slide that I showed you here,

  • there are at least two ways we are

  • using machine learning here.

  • The first, if you use Y Searcher--

  • so if you click that icon on the right for the mic

  • and you say something, there's machine

  • learning being used to understand what you're saying

  • and converting it to text.

  • So you say something.

  • That's really a bunch of bytes for the computer.

  • To convert that into actual English or whatever,

  • like you're speaking, it needs to understand that.

  • And there are pretty sophisticated machine

  • learning models that do that for you today.

  • Then the next part is once you have

  • that, what results to show you?

  • Google has billions or trillions of pages really indexed,

  • which you probably don't even think about.

  • What you see typically when you do that search

  • is just 10 pages on that first page.

  • And you really want those 10 pages

  • to be the best ones, the most appropriate ones for you.

  • In some cases, you might go to the next page or so.

  • But most people are just going to look at that first page

  • and make their decision.

  • So how do we really take those billions of pages,

  • match them to your query, and then even then

  • it's probably millions or maybe more,

  • and then put them down to those, whittle them down

  • to those top 10 pages?

  • And that's, again, machine learning

  • at work in many, many forms.

  • We're taking all these pages, understanding your query,

  • matching it through all the data that we have.

  • Maybe if you're looking for a restaurant near Mountain View,

  • we're going to leverage that and understand

  • that that needs to be combined with the kind of pages

  • that are there.

  • If it's something else, again, understanding

  • the query and the documents together, combining them,

  • and then ranking them, and sorting them,

  • that something that machine learning does really,

  • really well.

  • This example is of pictures.

  • We talked about vision models.

  • They're actually used in everything you do now.

  • So for example, over the years, image collections, the amount

  • of photos I have are huge.

  • They're just growing.

  • With digital cameras coming on, it's

  • been so much easier to take pictures and just

  • keep collecting them.

  • You never throw them away.

  • But then when you want to find a picture

  • and you really want to go through those collections,

  • it's really, really hard.

  • So how can you do that?

  • In this case, the same kind of research

  • that I was talking about, which understands those images,

  • is really indexing and labeling all the pictures

  • you have in Google Photos.

  • And in this particular example, if you

  • search for cherry blossom, it really

  • shows you all your art pictures of cherry blossom,

  • and can display them for you.

  • Another example is email.

  • So there is this thing called Smart Reply, which

  • we launched a few months ago, which

  • you might be using in email.

  • And what happens there is you get an email.

  • You typically, in this case, you probably

  • just want to get a quick, easy answer.

  • And it just gives you some choices.

  • Here, pick this one.

  • And what it's doing is it's really

  • understanding what your email thread is about

  • and then suggesting a bunch of replies ranking them

  • whatever makes the most sense and showing

  • them the most interesting ones to you if it makes sense.

  • Another one is Google Music.

  • There are many things happening in there.

  • And one of the things that we do is we recommend a lot of things

  • that you might want to play, maybe albums that you

  • might want to play, maybe actual songs

  • that you might want to play based on the time,

  • based on what you are trying to do right now,

  • based on your interest.

  • It's really understanding all of that, sorting them,

  • and, again, ranking them, recommending

  • the right things for you.

  • We only want to make it easy for you to find what you want.

  • The ideal case would be you go there.

  • Let's say it's afternoon or let's

  • say you're driving back from work.

  • And you typically like to listen to, say, classical songs

  • at that time.

  • It should really just recommend you

  • the classical song and maybe the top station

  • that you do every time.

  • You shouldn't have to look for it just because-- just

  • before this, you were doing something else.

  • So all of these are examples.

  • Machine learning can really make a difference

  • and are helping us in our everyday lives.

  • And one more that I'm going to talk about is spam.

  • So everybody uses email.

  • A few years ago, many 5, 10 years ago,

  • spam was this huge issue for everyone

  • where most of the emailing in our inbox

  • is starting to be spam.

  • You have to actually filter them out manually.

  • So it's not that from that time on to now spam is reduced.

  • In fact, spam continues to increase.

  • It's just there are filters, and algorithms

  • have gotten a lot better at recognizing what's spam.

  • Again, they're not perfect.

  • Occasionally you'll still see spam.

  • Or maybe we would say something is spam when it's not.

  • But they're getting so much better,

  • that I don't go to my spam for spam box

  • and see what I might have missed,

  • or the other way around.

  • I rarely get emails that's actually spam,

  • even though if you actually read the numbers about how much spam

  • emails there are, these are still

  • way more, maybe 10 times more than the real emails of spam

  • itself.

  • So lots of improvements in machine learning,

  • lots of improvements in the products

  • that are driven by machine learning.

  • So it's not about the algorithms or the technology itself.

  • It's about how do we apply it to the products that we build?

  • How do we make things better for all the things

  • that we're trying to do?

  • All of this technology, and algorithms

  • in the computer that we're building,

  • it's all towards solving real problems,

  • solving real needs for things that we do today.

  • So now let's talk a little bit about TensorFlow.

  • That's the product I need.

  • And it was built to help push machine learning forward.

  • Our team is a research team.

  • And part of the goal for us was to really push

  • machine learning forward.

  • And that's why we built it.

  • But that's not the angle, right?

  • We want to take that machine learning as well

  • and apply it like I was saying in all of these products.

  • TensorFlow allows us to do that, and now you as well.

  • And I'm going to tell you of how we can do this.

  • So it is open source.

  • It's a machine learning lab available on GitHub.

  • It's, in fact, the most popular machine

  • learning library we opened source six months ago.

  • And it's rapidly going to the top.

  • There's a lot of interest.

  • There are a lot of people using it in all kinds of areas.

  • Some of the research that I talked about, actually

  • people outside Google who've taken up TensorFlow, and used

  • it in doing a lot of those-- building those new things

  • and new ideas, people are including it

  • in different kinds of products as well.

  • So at the core, it's a library that's-- let's go a step

  • forward.

  • So what's this library for right there?

  • There are many different things that we can do.

  • Like I was saying, our group is doing a lot of research.

  • So it is for researchers who want to take machine learning

  • and really push it forward, make it much better than it is,

  • improve the algorithms, improve how we can do things with it.

  • So it is for them.

  • But then there are data scientists

  • who want to really take those algorithms-- let's

  • see our data, let's see our product.

  • And you want to improve that product

  • by playing this machine learning to the data that you have.

  • And so the data scientists can take those algorithms

  • because they're also open now.

  • They're made available to you on all kinds of places.

  • You can take those.

  • You can use those and make some improvements.

  • And then there are developers, like a lot of us

  • here, actually, who are trying to make, say, an app,

  • and want to make it better, or say, apply a product

  • to their data center.

  • And TensorFlow allows you to really take all the work

  • that researchers and data scientists have done,

  • and really run those same models, the train

  • models in production using the exact same APIs without having

  • to worry about-- OK.

  • This seems great, but how am I going

  • to deploy this to production.

  • That's another complexity that that's harder.

  • So we definitely try to make sure that from the beginning,

  • from the point that you get started in machine learning,

  • to actually pushing it out, to production,

  • it's a very short time cycle.

  • You shouldn't be stuck because of infrastructure.

  • That's taken care of you.

  • That's taken care of for you.

  • And you can focus on what's important for you

  • in the most ways, which is really building the products,

  • helping your customers, helping your users.

  • So just a couple examples of TensorFlow-- not going

  • to go too detailed into the code,

  • but this is a very simple example where we just

  • take two inputs, do a matrix multiply,

  • and just print out the result. So what it's doing

  • is in the first two cases, it's creating

  • two nodes in the graph.

  • And actually, let me just-- yeah.

  • It's creating two nodes in the graph,

  • and really doing a multiplication, which

  • is another node, which takes the inputs from the other two

  • nodes.

  • And then the last one really evaluates

  • based on the two inputs, combines them,

  • and generates the output.

  • This one's actually a regression problem.

  • So we talked about classification earlier.

  • In this case, it's trying to predict some values based on,

  • can we have some inputs?

  • Can we have some outputs?

  • You want to learn to predict the outputs.

  • It's basically doing something similar to what we were seeing

  • on the playground where you're taking inputs,

  • you're multiplying by matrix, and you haven't lost it.

  • You're trying to optimize it, which

  • is a typical machine learning thing.

  • And then you just loop through.

  • So the last, at the end, you basically have these.

  • You're going through a bunch of steps, and say 200 times

  • in this particular case, and you just print this out.

  • And over time you're basically learning

  • to predict things better.

  • Again, very simplistic examples, typical models

  • might be much bigger.

  • But in a lot of cases, you might also

  • be able to use models that are already there.

  • Why don't we just give you some examples

  • of the kind of performance we provide comparing it

  • to some standard benchmarks that are there on an open source

  • website.

  • And early on when we launched TensorFlow,

  • it was in the fastest library out there.

  • And there were a lot of improvements that we had to do.

  • But over the last six months, we've really focused on that.

  • And in this point, it's really at the top

  • in terms of compare compatible to everything else out there.

  • And we continue to improve it.

  • There's not the end of it.

  • There are many more improvements that you can do.

  • In this case, we're combating four different kinds of models.

  • In this case, they happen to be all image models.

  • And as you see, the load is actually better.

  • In a few cases, it's just slightly worse.

  • In few cases it's slight-- in one case

  • it's likely better, and so on.

  • So it's roughly about the same ballpark performance

  • at this point.

  • Now, one thing you can do with TensorFlow as well

  • is you can run it across many, many machines

  • rather then running on just one machine.

  • So the previous benchmark that I was showing you

  • was basically running on a single GPU

  • and training these models on GPUs, something

  • that a lot of people do.

  • In this particular case, we're training on many, many more

  • machines, so each machine having a single GPU

  • going all the way to 100.

  • And you can see it scales quite well.

  • In fact, up to 16 it's almost linear.

  • And then it's still going to be used to scale you in past that.

  • And this is something that we use internally as well

  • to train really large models and large data sets, which

  • is-- as the data is going, it's really important to be

  • able to do this, to be able to run on clusters, which many

  • of you probably do already.

  • Oops.

  • So next I'm going to talk a little bit about TensorFlow

  • for Poets.

  • So there's actually a code lot for this

  • as well if you want to look at it and look at it online,

  • or I believe there's a place here too.

  • So the idea here is-- let's say you have a bunch of images,

  • in this case flowers, and you want

  • to learn to recognize those flowers,

  • or maybe you just identify if it's a flower

  • or if these are five different kinds of flowers.

  • And what you can do is you don't have

  • to have a million examples.

  • You might just have 1,000 examples, lets say.

  • But you can get started from a model that works really,

  • really well.

  • For example, our best image model

  • that you've open sourced and made available,

  • and that's what this example talks about.

  • Take the top part of it, just train that

  • on your data set from the point that it's already trained on.

  • So it's just going to learn to classify the flowers.

  • It already understands image well,

  • so it doesn't have to really know what kind of features

  • are important for classifying images.

  • It knows that already.

  • It can just take your particular models, your particular data,

  • and learn to just separate those out into different classes.

  • So there are a few tools that I would

  • like to mention at the end that you can play with,

  • or that you can use before you go in terms of just making

  • things work well for you.

  • So a lot of you are probably familiar with Apache Spark,

  • which is used for doing analysis and doing

  • learning on large data sets.

  • So there is an integration of Spark

  • with TensorFlow-- somebody at Databricks

  • did that-- where you can actually take some images.

  • You can create.

  • It's very simplistic in the sense

  • you can create a lot of different models

  • and really use it to pick the best model.

  • You do that a lot because there are some having parameters

  • like, say, learning rate or how big should your model be,

  • et cetera.

  • You maybe try 10 different models and use

  • that to pick the best one.

  • Another thing I would like to point you do

  • is Google's own Cloud Machine Learning Platform.

  • We announced this a couple of months ago.

  • It's under limited preview right now, but you can sign up,

  • and we will make it accessible to you

  • soon, where you can really have the same power of TensorFlow

  • that you have on your machine say,

  • but maybe don't have the compute resources,

  • or you don't want to manage all the resources that you need

  • to really run at large scale.

  • And cloud takes care of all of that for you.

  • So you can experiment on your machine,

  • make sure you understand what you want to do, and then

  • push it off to the cloud, let it take care

  • of running large scale stuff.

  • There are a few examples of what you

  • can do with models that have already been trained by others

  • and really take them and just apply them.

  • There are some APIs on the forum today

  • for what's called the Mobile Mission APIs, which

  • allows you to do face tracking.

  • So let's say your pictures, it allows you to-- it

  • identifies parts of the faces.

  • It first identifies the face, and then

  • parts of the face like the eyes, nose, et cetera,

  • so you can do different kinds of things with it.

  • It also recognizes barcodes and QR codes

  • and automatically understands and parses them for you.

  • On the cloud there are also some APIs for vision and speech

  • in translation, which you can basically just send your image,

  • and it tells you what kind of image

  • it is among a lot of different classes.

  • For speech, again, you can upload data, upload speech,

  • and it will give you back text for whatever kind.

  • For translation, you can give it text in one language,

  • say English or Japanese or whatever makes sense for you,

  • and it's going to convert that to another language.

  • So many different APIs, and this is a growing area

  • that you are going to need to watch out for.

  • So now for the future.

  • It is really.

  • I started by saying it's really not the future, but it is.

  • It i the present and the future.

  • I just wanted to call out that it's important for you

  • to start thinking about it now and not wait for the future,

  • really.

  • And it's really your turn to start

  • thinking about how you can incorporate machine learning

  • into everything that you do.

  • What people are looking for today

  • is not just applications that do what they want.

  • There are tons of things that are already doing that.

  • How can you add that edge?

  • How can you how you make it really better for that user?

  • By understanding what they want before they

  • ask for every single thing.

  • And that's really something that you can do.

  • Think about this as you build your next application,

  • as you build your next project.

  • And I think that's how we'll change the future.

  • That's how we'll make machine learning part of the future.

  • So I wanted to leave you with some place with some pointers

  • to where you can read more about TensorFlow.

  • It's on GitHub, and their docs are tensorflow.org.

  • And going back to my reference for Tik-Tok,

  • Tik-Tok is waiting for you to really make the difference

  • to do these things now.

  • And thank you.

  • [APPLAUSE]

  • [MUSIC PLAYING]

Cool.

字幕と単語

ワンタップで英和辞典検索 単語をクリックすると、意味が表示されます

A2 初級

機械学習は未来ではない - Google I/O 2016 (Machine learning is not the future - Google I/O 2016)

  • 115 10
    Amy.Lin に公開 2021 年 01 月 14 日
動画の中の単語