Placeholder Image

字幕表 動画を再生する

  • >>

  • So it's my pleasure to introduce to you Geoff Hinton, who is a pioneer in machine learning

  • and neural nets, and more recently [INDISTINCT] architectures. And then, I think that's going

  • to be topic of today. So take it over. >> HINTON: Okay. So I gave it to Okeer a couple

  • of years ago. And the first 10 minutes or so will be an overview of what I said there,

  • and then I'll talk about the new stuff. The new stuff consists of a better learning module.

  • It allows you to learn better in all sorts of different things, like, learning how images

  • transform, learning how people walk, and learning object recognition. So the basic learning

  • module consists of some variables that represent things like pixels, and these will be binary

  • variables for that. Some variables that represent--these are latent variables, they're also going to

  • be binary. And there's a bipartite connectivity, so these guys are connected to each other.

  • And that makes it very easy if I give you the states of the visible variables to infer

  • the states of the hidden variables. They're all independent given the visible variables

  • because it's a non-directed graph. And the input procedure just says, the probability

  • of turning on hidden unit "hj" given this visible vector "v" is the logistic function

  • of the total what he gets from his audience, so very simple for the hidden variables. Given

  • the hidden variables, we can also infer the visible variables very simply. And if we want

  • some--if we put some weights on the connections and we want to know what this model believes,

  • we can just go back and then forward inferring all the hidden variables in parallel than

  • all the visible ones. Do that for a long a time, and then you'll see examples of the

  • kinds of things it likes to believe. And the end of learning is going to be to get it to

  • like to believe the kinds of things that actually happen. So this thing is governed by an energy

  • function that has given the weights on the connections. The energy of a visible plus

  • a hidden vector is the sum overall connections of the weight if both the visible and hidden

  • units are active. So I'm going to pick some of the features that are active, you adding

  • the weight, and if it's a big positive weight, that's low energy, which is good. So, it's

  • a happy network. This has nice derivatives. If you differentiate it with respect to the

  • weights, you get this product of the visible and hidden activity. And so, that the derivative

  • is going to show up a lot in the learning because that derivative is how you change

  • the energy of a combined configuration of visible and hidden units. The probability

  • of a combined configuration, given the energy function, is E to the minus the energy of

  • that combined configuration normalized by the partition function. And if you want to

  • know the probability of a particular visible vector, you have to sum all the hidden vectors

  • that might go with it and that's the probability of visible vector. If you want to change the

  • weights to make this probability higher, you always need to lower the energies of combinations

  • of visible vector on the hidden vector that would like to go with it and raise the energies

  • of all other combinations, so you decrease the computation. The correct maximum likelihood

  • learning rule that is if I want to change the weights so as to increase the log probability,

  • that this network would generate the vector "v" when I let it just sort of fantasize the

  • things it like to believe in is a nice simple form. It's just the difference of two correlations.

  • So even though it depends on all the other weights, it shows up this is difference of

  • correlations. And what you do is you take your data, you activate the hidden units,

  • that's to classify the units, and then we construct, activate, we construct activate.

  • So this is a mark of chain. You run it for a long time, so you forgot where you started.

  • And then you measure the correlation there, start with the correlation here. And what

  • you're really doing is saying, "By changing the weights in proportion to that, I'm lowering

  • the energy of this visible vector with whatever hidden vector it chose. By doing the opposite

  • here, I'm raising the energy, the things I fantasize." And so, what I'm trying to do

  • is believe in the data and not believe in what the model believes in. Eventually, this

  • correlation will be the same as that one. In most case, nothing will happen because

  • it will believe in the data. In terms that you can get a much quicker learning algorithm

  • where you can just go on and [INDISTINCT] again, and you take this difference of correlations.

  • Justifying that is hard but the main justification is it works and it's quick. The reason this

  • module is interesting, the main reason it's interesting is you can stack them up. That

  • is for accompanied reason you're not going to go into it, it works very well to train

  • the module then take that activities of the feature detectors, treat them as so they were

  • data, and train another module on top of that. So the first module is trying to model what's

  • going on in the pixels by using these feature detectors. And the feature detectors would

  • tend to be highly correlated. The second model is trying to model a correlation among feature

  • detectors. And you can guarantee that if you do that right, every time you go up a level,

  • you get a better model of the data. Actually, you can guarantee that the first time you

  • go up a level. For further levels, all you can guarantee is that there's a bound on how

  • good your model of the data is. And every time we add another level, that bound improves

  • if we had it right. Having got this guarantee that something good is happening as we add

  • more levels, we then violate all the conditions of mathematics and just add more levels in

  • sort of [INDISTINCT] way because we know good things are going to happen and then we justify

  • by the fact that good things do happen. This allows us to learn many lesser feature detectors

  • entirely unsupervised just to model instruction of the data. Once we've done that, you can't

  • get that accepted in a machine learning conference because you have to do discrimination to be

  • accepted in a machine learning conference. So once you've done that, you add some decision

  • units to the top and you learn the connections discriminatively between the top-layer features

  • and the decision units, and then if you want you can go back and fine-tune all of the connections

  • using backpropagation. That overcomes the limit of backpropagation which is there's

  • not much information in the label and it can only learn on label data. These things can

  • learn on large amounts of unlabeled data. After they've learned, they you add these

  • units at the top and backpropagate from this small amount of label data, and that's not

  • designing the feature detectors anymore. As you probably know at Google, designing feature

  • detectors is the art of things and you'd like to design feature detectors based on what's

  • in the data, not based on having to produce labeled data. So the edge of backpropagation

  • was design your feature detectors so you're good at getting the right answer. The idea

  • here is design your feature detectors to be good at modeling whatever is going on in the

  • data. Once you've done that, just have a so slightly fine-tune and so you better get right

  • answer. But don't try and use the answer to design feature detectors. And Yoshua Bengio's

  • lab has done lots of work showing that this gives you better minima than just doing backpropagation.

  • And what's more minima in completing different part of the space? So just to summarize this

  • section, I think this is the most important slide in the talk because it says, "What's

  • wrong with million machine learning up to a few years ago." What people in machine learning

  • would try to do is learn the mapping from an image to a label. And now, it would be

  • a fine thing to do if you felt that images and labels are rows in the following way.

  • The stuff and it gives rise to images and then the images give rise to the labels. Given

  • the image, the labels don't depend on the stuff. But you don't really believe that.

  • You only believe that if a label is something like the parity of the pixels in the image.

  • What you really believe is the stuff that gives rise to images and then the labels that

  • goes with images because of the stuff not because of the image. So it's a cow in a field

  • and you say cow. Now, if I just say cow to you, you don't know whether the cow is brown

  • or black, or upright or dead, or far way. If I show an image of the cow, you know all

  • those things. So this is a very high bandwidth path, this is a very low bandwidth path. On

  • the right way to associate labels with images is to first learn to invert this high bandwidth

  • path. And we can currently do that because vision works basically. The first store you

  • look at then, you see things. And it's not like it might be a cow, it might be an elephant,

  • it might be electric theater. Basically, you get it right nearly all the time. And so we

  • can invert that pathway. Having learned to do that, we can then learn what things are

  • called. But you get the concept of a cow not from the name, but from seeing what's going

  • on in the world. And that's what we're doing and then later as I say from the label. Now,

  • I need to do one slight modification to the basic module which is I had binary units as

  • the observables. Now, we want to have linear units with Gaussian noise. So we just change

  • the energy function of it. And the energy now says, "I got a kind of parabolic containment

  • here." Each of these linear visible units has a bias which is like its mean. And it

  • would like to sit here and moving away from that [INDISTINCT] energy. The parabola is

  • the negative log of the Gaussian [INDISTINCT]. And then the input that comes from the hidden

  • units, this is just vi, hj, wij, but Vs have to be scaled by the standard deviation of

  • the Gaussian there. If I ask--if I differentiate that with respect to a visible activity, then

  • what I get is hj, wij divided by the sigma I. And that's like an energy gradient. And

  • what the visible unit does when you reconstruct is it tries to compromise between wanting

  • to sit around here and wanting to satisfy this energy gradient, so it goes to the place

  • where this two gradients [INDISTINCT] opposite and you have--that's the most likely value

  • and then you [INDISTINCT] there. So with that small modification we can now deal with real

  • value data with binary latent variables and we have an efficient learning algorithm that's

  • an approximation of [INDISTINCT]. And so we can apply it to something. So it's a nice

  • speech recognition task that's been well organized by the speech people where there's an old

  • database called TIMIT, it's got a very well-defined task for phone recognition where what you

  • have to do is you're given the short window speech, you have to predict the distribution,

  • the probability for the central frame of the various different phones. Actually, each phone

  • is modeled by 3-state HMMs, sort of beginning middle and end, so you have to predict for

  • each frame is it the beginning middle or end of each with the possible phones, there's

  • a 183 of those things. If you give it a good distribution there to sort of focus on the

  • right thing then all the post-processing will give you back where the phoning bandwidth

  • should be and what your phone arrow radius, and that's all very standard. Some people

  • use tri-phone models. We're using bi-phone models which aren't quite as powerful. So

  • now we can test high goodwill by taking 11 frame of speech. It's 10 milliseconds per

  • frame but each frame is looking at like 25 milliseconds of speech and predicting the

  • phone at the middle frame. We use the standard speech representation which is mel-cepstral

  • coefficients. There's 30 of those, and there are differences and differences and difference,

  • differences; and we feed them in to one of these deep nets so. So here's your input,

  • 11 frames and 39 coefficients. And then--I was away when the student did this and he

  • actually believed what I said. So he thought adding lots and lots of hidden units was a

  • good idea. I've started it too. But he added lots of hidden units all unsupervised, so

  • all this green connections are learned without any use of the labels. He used to bottleneck

  • there, so the number of Reg connections will be relatively small. These are not--these

  • have to be learned using discriminative information. And now you're back propagating the correct

  • answers through this whole net for about a day on a GPU board or a month on a core, and

  • it does very well. That is the best phone error rate we got was 23%. But the important

  • thing is whatever configuration you use, how many hidden layers as long as they are plenty

  • and whatever widths and whether you use this bottleneck or not, it gets between 23% and

  • 24%. So it's very robust to the exact details of how many layers and how wide they are.

  • On the best previous result on TIMIT for things that didn't use speaker adaptation is 24.4%

  • and that was averaging together on lots of models, so this is good.

  • >> So each of these layers that's four million weights?

  • >> HINTON: Yup, four million weights. So we're only training one, two, three, one, two, three,

  • we're training, you know, about 20 million weights. Twenty million weights is about 2%

  • of a cubic millimeter of cortex. I think so, this is a tiny brain. The last, probably,

  • all you need for [INDISTINCT] recognition. >> Why did they start with the differences

  • and double differences of the MFCCs that you're going into a thing that could learn to do

  • that itself if they wanted to? >> HINTON: That's a very good question 'cause

  • you are sitting at the end. It's an extremely good question because the reason I put the

  • differences and double differences is so they can model a data with a diagonal co-variance

  • metric--diagonal co-variance model--and you can't model the fact that overtime two things

  • turn to be very much the same where modeling co-variance is, unless you actually put the

  • differences into the data and you model the differences directly. So it allows you to

  • use a model that conquered with co-variances. Later on we're going to show a model that

  • conquered with co-variances and then we are going to do what the client always said you

  • should do, which is throw away the mel-cepstral representation and use the better representation

  • in speech. >> I said that?

  • >> HINTON: Yes, you said that to me the last time [INDISTINCT].

  • >> Smart guy. >> HINTON: Okay, so the new idea is to use

  • a better kind of module. This module already works pretty well, right? You know, it does

  • well at forming recognition, it does well in all sort of things. It can't model multiplicative

  • interactions very well. It can model anything with enough training data, but its not happy

  • modeling multiplies. It multiplies all over the place. I'll show you a bunch of place

  • where you need multiplies. Here's the sort of main example of why you need multiplies.

  • Supposed I want to, from a high level of description of an object the name of the shape and it's

  • pose, size, position orientation. So, first of all, I want to generate the parts of an

  • object and I want them to be related correctly to each other. I could use very accurately

  • top down model that says none of this square and none pose gram, because I generate each

  • piece in exactly the right position; that would require high bandwidth. Or I could be

  • sloppy and I could say, "I need to generate this side, not sort of a representation and

  • distribution of where this side might be. And I'll generate corners and other sides

  • and they're all a bit sloppy. And if I picked one thing from each distribution, it would

  • make a nice square. But I could also top-down specify how these things should be pieced

  • together. In effect, I can specify a macro form of field; this is what goes with what?

  • And then I can clean this up knowing these distributions and pick a square like that.

  • Of course, I might sometimes pick a square that has slightly different orientation or

  • slightly different size, but it'll be a nice clean square because I know how they go together.

  • And so that's a much more powerful kind of generation model, and that's what we want

  • to learn to do, and so we are going to need hidden units up here to specify interactions

  • between visible units here, as opposed to just specifying input of visible units. There's

  • an analogy for this, which is, if I'm an officer and there's a bunch of soldiers and I want

  • them to stand in the square, I could get out my GPS and I can say, "Soldier number one,

  • stand at this GPS coordinates. And soldier number two, stand at these GPS coordinates."

  • Now, if I use enough digits, I'll get a nice neat rectangle, or I could say, "Soldier number

  • one, stand roughly around here. And then soldier number two, hold your arm out and stand this

  • distance from soldier number one." And that's a much better way to get a neat rectangle.

  • It will cause far less communication. So what you're doing is you're downloading roughly

  • where people should stand and then how they should relate to each other. We have to specify

  • the relations not just where they should be. And that's what we'd like in a powerful [INDISTINCT]

  • model. So, we're going to aim to get units in one layer to say how units in the layer

  • below should latterly interact when you generate it. It's going to turn out you don't need

  • to worry about these lateral interactions when you recognize it. When you generate,

  • you do. To do that, we're going to need things called third-order Boltzmann machines, which

  • has three lane tractions. So Terry Sejnowski, he pointed out a long time ago that we have

  • an energy function like this where this was V and this was H, but these are just binary

  • variables. And we could probably well write down an energy function like this with three

  • things in tract, then we have a three layer weight. And if you think about these three

  • things now, K, the state of K is acting like a switch. When K is on, you effectively have

  • this weight between I and J. When K is off, this weight disappears. And it happens every

  • which way because it's symmetric. So using an energy function like this, we can allow

  • one thing to specify how two other things should interact. So each hidden unit can specify

  • a whole mark of random field over the pixels if you want. But that sort of begins to make

  • you worry because a mark of random field has a lot of parameters in it. And if you start

  • counting in the sits here, if you have any of these and any of those and none of those,

  • you get enqueued to these parameters which is rather a lot. If you're willing to use

  • enqueued parameters, you can now make networks to look like this. Suppose I have two images

  • and I want to model how images transform over time, unless suppose I'm just moving random

  • dots around, have a pan of random dots and I translate it. Well, if I see that dot and

  • I see that dot, that's some evidence for a particular translation. And so if I put a

  • big positive weight there, this triangle is meant to interrupt that big three-way weight.

  • Then when this and this around, they'll say it's very good to have this guy up. It would

  • have been nice at low energy state. If I also see this pair of dots, I'll get more routes,

  • though, and this guy should be--and I will turn this guy on. If however this pixel went

  • to here, I'll go for this guy. And if this pixel also went to there but this guy--so

  • these guys are going to represent coherent translations of the image, and it's going

  • to be able to use these three-way weights to take two images and extract two units that

  • represent the coherent translation. It'll also be able to take the pre-image and the

  • translation, and compute with pixel should be on here. Now what we're going to do is

  • take that basic model and we're going to factorize it. We're going to say, "I've got these three-way

  • weights and I've got too many of them." So, I'm going to represent each three-way weight

  • as the product of three two-way things. I'm going to introduce these factors and each

  • factor is going to have these many parameters which is just pro-factor, is just a linear

  • number of parameters. If I have about N factors, I end up with only N square of these weights.

  • And if you think about how pixels transform in a new image, they don't do random permutations.

  • It's not that this pixel goes on that one, goes here. Pixels do sort of consistent things,

  • so I don't really need enqueued parameters because I'm just trying to model these fairly

  • consistent transformations, which is a limited number, and I should be able, too, in many

  • less parameters. And this is the way to do it. So, that's going to be our new energy

  • function given the bias terms. One way of thinking about how modeling a weight is I

  • want these tensor of three-way weights. If I take an IF product and two vectors like

  • this, I'll get a matrix that has rank one. So I get a three-way product. I'll get a tensor

  • that has rank one. And if I now add up a bunch of tensors like that so each factor now, each

  • F, specifies a rank one tensor, by adding up a bunch of them, I can model any tensor

  • I like if I use N square factors. If I use any N factors, I can model most regular tensors

  • but I can't model arbitrary permutations, and that's what we want. If you ask how does

  • inference works now, inference is still very simple in this model. So here's a factor.

  • Here's the weights connecting it to, say, the pre-image. Here's the weights connecting

  • it to the post-image. Here's the weights connecting it to the hidden units. And to do inference

  • what I do is this. Supposed I only have that one factor. I would multiply the pixels by

  • these weights; add all that up so I get a sum of this vertex. I do the same here; I

  • get a sum of this vertex. Then I multiply these two sums together to get a message going

  • in the center of the hidden units. And as that message goes to the hidden unit, I multiply

  • it by the weight on my connection. And so what the hidden unit will see is this weight,

  • turns the product of these two sums, and that is the derivative of the energy with respect

  • to the state of this hidden unit, which is what it needs to know to decide whether to

  • be on or off, it wants to go into what other state below the image. And all the hidden

  • units remain independent even though I've got these multipliers now. So this is much

  • better than putting in another stochastic binary unit here. If I put a stochastic binary

  • unit in here, the hidden units would cease to be independent and inference will get tough.

  • But this way, whether a deterministic factor that's taking a part of these two sums, inference

  • remains easy. The learning also remains easy. So this is the message that goes from factor

  • F to hidden unit H, and that message is the product that we got of these two lower vertices;

  • the product of the sums, the compute on the pre-image and the post-image. And the way

  • you learn the weight on the connection from factor F to hidden unit H is by changing the

  • weight so as to lower the energy when you're looking at data, and raise the energy when

  • you're constructing these from the model or just reconstructing things from the hidden

  • units you got from data. And those energy groups, they just look like this. They're

  • just the product of the state of the hidden unit and the message that goes to it when

  • you're looking at data and the state of the hidden unit and the message that goes through

  • it when you're looking at samples from the model or reconstructions. So it's still a

  • nice pair-wise learning rule. So everything is pair-wise still, so you might fit into

  • the brain. Now, if we look what one of these factors does when I show random dot patterns

  • to translate, then we can look at the weights connecting it to the pre-image, and that's

  • a pattern of weights where white is a big positive weight, black is a big negative weight

  • because that would have a learned a great in connecting it to the pre-image and this

  • will have learned a great in connecting it to the post-image. With a hundred factors,

  • I'll show you what Roland learned. So, those are the hundred the hundred factors connecting--these

  • are the receptive fields at the factors in the pre-image. And remember it's looking at

  • translating dots, and these are the factors in the post-image. And you see, it's basically

  • learned the freer basis and it's learned to translate things by about 90 degrees. And

  • that's a very good way of handling translation. Mathematicians say things like, "The freer

  • basis is a natural basis for modeling translation." I don't really know what that means, but just

  • learn the freer basis on that. And if you get rotations, it learned the different basis.

  • So this is the basis that learns for rotations. You see it learns about yin and yang here.

  • Oops [INDISTINCT]. Okay, that's the basis for rotations. One other thing you could do

  • is train it just on single dot pans and translating in a coherent way and then test it on two

  • overlaid dot patterns but they're translating different directions. It's never seen that

  • before. It's only been trained on coherent motion where we're going to test it on what's

  • called transparent motion. In order to see what it thinks, when we train the uncivilized,

  • there's no labels anywhere, we never tell it what the notions are, we need some way

  • to seeing what it's thinking, so we add a second hidden layer that looks at the hidden

  • units representing transformations and it's fairly sparse. So the units on that second

  • hidden layer will be tuned to particular directions of motion. And then to see what it's thinking,

  • we take the directions those units like weighted by how active those units are and I will tell

  • you what directions it thinks it's seeing. Now when you show it transparent motion and

  • you look at those units in the second hidden layer, if the two motions are within about

  • 30 degrees, it sees a single motion of the average direction. If they're beyond about

  • 30 degrees, it sees two different motions and once more they're repelled from each other.

  • That's exactly when I was with people, and so this is exactly how the brain works. Okay.

  • There's going to be a lot of that kind of reasoning in this talk. I'm going to on to

  • time series models now. So, we'd like to model not just static images, for example, we like

  • to model video. To be [INDISTINCT] we're going to try something a bit simpler. When people

  • do time series models, you would nearly always like to have a distributed non-linear representation,

  • but that's hard to learn. So people tend to do dumb things like Hidden Mark up Models

  • or Linear Dynamical Systems which either give up on the distributed or on the non-linear,

  • but are easy to doing inference. What we're going to come up with is something that has

  • the distributed and the non-linear and is easy to do inference, but the linear algorithm

  • isn't quite right but it's good enough. It's just an approximation to make some [INDISTINCT].

  • And the inference also is ignoring the future and just basing things on the past. So, here's

  • a basic module, and this is with just two-way interactions. This is the Restricted Boltzmann

  • Machine with visible units and hidden units. Here are the previous visible frames. These

  • are all going to be linear units. And so, these blue connections are conditioning the

  • current visible values on previous observed values in a linear way. So, it's called an

  • autoregressive model. The hidden units here are going to be binary hidden units; they're

  • also conditioned on previous visible frames, and learning is easy in this model. What you

  • do is you take your observed data, and then given the current visible frame and given

  • the previous visible frames, you got import to the hidden units, they're all independent

  • given the data, so you can separately decide what states they should be in. Once you fixed

  • states for them, you now reconstruct the current frame using the input you're getting from

  • previous frames and using the top you got in from the hidden units. After we construct,

  • you then activate the hidden units again. When you say the difference in the power statistics

  • with data here and the reconstructions here to learn these weights and you take the difference

  • on activities of these guys with data with reconstructions to get signal that you can

  • used to learn these weights or these weights. So learning is straightforward and it just

  • depends on differences, and you can learn a model like this. After you've learned it,

  • you can generate from the model by taking some previous frames. These inputs, the conditioning

  • inputs, in effect, fixed the biases of these to depend on the previous frames. So, these

  • are the dynamic biases, and with these biases fixed, you just get backwards and forwards

  • for awhile and then pick a frame there, and that's your next frame you regenerated, then

  • you keep going. So, we can generate from the model once it learns so we can see what it

  • believes. >> You always go back two steps in time or

  • is that just an example? >> HINTON: Sorry.

  • >> Oh, you were just going back only two steps in time?

  • >> HINTON: No, we're going to get back more sets in time.

  • >> Okay, and you let... >> HINTON: I just got lazy with the PowerPoint.

  • Now, one direction we could go from here is to higher level models. That is, having learned

  • this model where these hidden units are all independent given the data, we could say--well,

  • what I've done is I've turned the visible frames into the hidden frames map. And it

  • turns out you can get a better model if you take these hidden frames, a model what's going

  • here, and now you put in conditioning connections between the hidden frames and more hidden

  • units that don't have conditioning here that don't interact with other hidden unit. [INDISTINCT]

  • in this model. Then you can prove that if you do this right, then you'll get a better

  • model of the original sequences or your improver band on the model of the original sequences.

  • So you can [INDISTINCT] lots of layers like that. And when you have more layers, it generates

  • better. But I'm going to go in a different direction. I'm going to show you how to do

  • it with three-way connections. And we're going to apply it to motion-capture data, so you

  • put reflective markers on the joints, you have lots of infrared cameras, you figure

  • out where the joints are in space. You know the shape of the body so you go backwards

  • through that to figure out the joint angles and then the frame of data is going to consist

  • of 50 numbers, about 50 numbers which are joint angles and the translations and rotations

  • of the base of the spine. Okay. So, imagine we got--one of these mannequins you see in

  • art shop windows, we got a pins stuck in the base of his spine and we can move him around

  • and rotate him using this pin and we can also wiggle his legs and arms. Okay. And what we

  • want him to do is as we move him around, we want him to wiggle his legs and arms so his

  • foot appears to be stationary on the ground and he appears to be walking. And he'd better

  • wiggle his leg just right as we translate his pelvis, otherwise his foot will appear

  • to skid on the ground. And we're going to model him, we can do hierarchal model like

  • I just showed you or we can a three-way model like this where we condition six earlier of

  • frames, this is a current visible frame, here's basic bolts in the machine accept that it's

  • neither one of these 3-way things where these are factors. And we have a 1-of-N style variable.

  • So, we have data and we tell it the style when we're training it, so that's sort of

  • semi-supervise. It learns to convert that 1-of-N representation in to a bunch of real

  • value features and then it uses this real value features as one of the inputs to a factor.

  • And what the factor are really doing is saying, these real value features are modulating the

  • weight matrixes that use for conditioning and also this weight matrix that use in your

  • parallel linear model. So, these are modulating an auto aggressive model. That's very different

  • from switching between auto aggressive model it's much more powerful. Yeah?

  • >> I missed with what this one event is...? >> HINTON: So, we're going to have data of

  • someone walking in various different styles. >> Styles of walking.

  • >> HINTON: The style of walking. Yeah. >> So you mean your earlier diagram when you

  • can't beat history, it looked like there was nothing to keep track in the relative order

  • of the earlier direct delta direct link because... >> HINTON: Yes.

  • >> ... is there anything in the model that cares about that relative...?

  • >> HINTON: Yeah. Yeah. The weights on the connections will tell you which frame it's

  • coming from. Right. In the earlier model, there were two blue lines, they're different

  • matrixes and they have different weights on. >> There's nothing from two steps to previews

  • to ones step previews, right, it just skip all the way?

  • >> HINTON: It just skipped all the way, right. You just...

  • >> Will that continue to happen? >> HINTON: Yes. In other words there's direct

  • connections from all six previous frames to the current frame for determining the current

  • frame. >> Right. And then what links from the six

  • frames of the fifth earliest report? >> HINTON: Well, there where when you were

  • computing what the fifth frame was doing, right?

  • >> Okay. >> HINTON: But when we're computing this frame

  • we have direct connections from it. Okay. So, we're now going to train this model, it's

  • relatively easy to train especially on the GPU board, and then we're going to generate

  • from it, so we can see sort of what it learned and we can judge if it's doing well by whether

  • the feet slip on the ground. All right. >> [INDISTINCT]

  • >> HINTON: We'll get there. >> Sorry.

  • >> HINTON: Here's a normal walk. Maybe, at least they're willing. Okay. So I was generating

  • from the model--he's deciding which direction to turn in, and he's deciding, you know, he

  • needs to make the outside leg go farther than the inside leg and so on. If we--we have one

  • model but if we flip the style label to say, gangly teenager, he definitely looks awkward.

  • Right. We've all been there. I think this is a computer science student. My main reason

  • for thinking that is if you asked him do a graceful walk, it looks like this. And that's

  • definitely C3PO. >> [INDISTINCT].

  • >> HINTON: Now, I think this was a student [INDISTINCT]--but he's very good. You can

  • ask him to walk softly like a cat. We're asking to model at present, right? The model looks

  • pretty much like the real data the real data obviously the feet are planted better but

  • notice, he can slow down then speed up again. Auto aggressive models can't do things like

  • that, Auto aggressive models have a biggest size in value, the size is bigger than one

  • in which case they explode or a smaller one in which case they die and the way to keep

  • them alive is by keep--you keep injecting random noise so that they stay alive and that's

  • like making a horse walk by taking a dead horse and jiggling it, it's kind of--it's

  • not good. Now, he doesn't have any model of the physics so, in order to do this kinds

  • of stumbles, there had to be stumble similar to that in the data but when he stop in which

  • he stumble he did when, he's entirely determining. We could make him do a sexy walk but you're

  • probably not interested in that. >> I just order a chicken.

  • >> HINTON: You want dinosaur the chicken? Where's dinosaur the chicken?

  • >> And chicken, number five. >> At number five.

  • >> HINTON: Oh, no, that's dinosaur and chicken. That's a blend. Maybe a switch. He's got quite a lot of foot

  • [INDISTINCT] that's probably a blend. This is doing a sexy walk and then you flip the

  • label to normal and then you flip it back to sexy. It's never seen any transitions but

  • because all one model, it can do reasonable for transitions.

  • >> So you have these hundred style variables, can you de-couple those form the one event

  • style and just make up new styles by playing with those...

  • >> HINTON: Yup. Yup. Now, you can also give it many more labels when you train, you can

  • give it speed, stride length all sorts of things then you can control it very well,

  • yeah. Okay. So, you can learn time series at least for 50 dimensional data and obviously

  • what we all want to do is apply that to video but we haven't done that yet. Except for some

  • very simple cases. The last thing I'm going to show is the most complicated use of these

  • 3-way models. One way of thinking of it, so that it's similar to the previous uses, is

  • that we take an image and we make two copies of it but they have to be same. And then we

  • insist the weights that go from a factor of this copy are the same as the weights that

  • go from the factor of this copy. So if I=J, WI=WJF. Inference is still easy in fact inference

  • here will consist of--you take these pixels times these weights to get a weighted sum,

  • and then you square it because this is going to be the same weighted sum. So, inference

  • consist--take linear filter, square its output, and send it by these weights to the hidden

  • units. That's exactly the model called the [INDISTINCT] energy model, is right, kind

  • of linear filter. This is being proposed both by Vision people by Adelson and Bergen, a

  • long time ago in the '80s, and by neuroscientists. So, neuroscientists had tried to take simple

  • cells, my point vaguely about, and look at what polynomial they're output is of their

  • input, and Yang Dan at Berkeley says it's between 1.7 and 2.3, and that's means two.

  • So, this looks quite like models that were proposed for quite different reasons and it

  • just drops out of taking a 3-way imaging model and factorizing it. The advantage we have

  • is that we have a learning algorithm for all these weighs now, when we generative model.

  • So now we can model covariances between pixels, and the reason that's good is–-well, here's

  • one reason why it's good. Suppose I asked you to define a vertical edge. Most people

  • will say, "Well, vertical edge is something that light on the side and dark on that side.

  • Well no, maybe it's light on this side and dark on that side, but you know. Well, it

  • could be light up here and dark down there, and dark up here and light down there." Okay.

  • Or it could be texture edge. It's getting--oh, it might actually be a disparity edge. Well,

  • the manner should be motion this side and no motion that side. That's a vertical edge

  • too. So, a vertical edge is a big assortment of things, and what all those things have

  • in common is vertical edge is something where you shouldn't do horizontal interpolation.

  • Generally new image, horizontal interpolation works really well. A pixel is the average

  • of its right and left neighbors, pretty accurately almost all the time. Occasionally it breakdowns,

  • and the place it breaks down is where there is a vertical edge. So, a real abstract definition

  • of vertical edge is breakdown of horizontal interpolation. And that's what our models

  • are going to do. A hidden unit is going to be putting in interpolation, and it's actually

  • going to turn-off sort of reverse logic, when that breaks down its going to turn-off, so

  • one way of seeing it is this. If this hidden unit here is on, it puts in a weigh between

  • pixel I and pixel J that's equal to this weight times this weight, times this weight. Okay.

  • Since theseokay, that's good enough. So, these are controlling affectively the mark

  • of random field between the pixels, so we can model covariances nicely. Because the

  • hidden units are creating correlations between the visible units reconstruction is now more

  • difficult. We could reconstruct one image given the other image, like we did with motion,

  • but if you want to reconstruct them both and make them identical it gets to be harder.

  • So, we have to use a different mathical Hybrid Monte Carlo. Essentially you start where the

  • data was and let it wonder away from where it was but keeping both images the same. And

  • I'm not going to go to Hybrid Monte Carlo, but it works just fine for doing the learning.

  • And the Hybrid Monte Carlo is used just to get the reconstructions and the learning algorithm

  • just the same as before. And what we're going to do is we're going to have some hidden units

  • that are using these 3-way interactions to model covariances between pixels, and other

  • hidden units are just modeling the means. And so we call-–for meaning covariance,

  • we call this mcRBM. Here's an example of what happens after it's learned on black and white

  • images. Here's an image patch. Here's its reconstruction of the image patch, if you

  • don't have noise, which is very good, from the mean and covariance units. Here's the

  • stochastic reconstruction which is also pretty good. But now we're going to do something

  • funny, we're going to take the activations of the covariance units. The things that are

  • modeling which pixels are the same with which other pixels and we're going to keep those.

  • But we are going to take the activations of the mean units, so we're going to throw those

  • away, and pretend that the means from the pixels look like this. Well, let's take this

  • one first. We tell all the pixels have the same value, except these are which are much

  • darker and it now tries to make that information about means fit in with these information

  • about covariances which is of these guys should be the same but very different from these

  • guys. And so, it comes up with the reconstruction that it looks like that. Where you see it's

  • taken this dark stuff and blurred across this region here. If we just give it four dots

  • like that, and the covariance matrix you've got from there, it'll blur those dots out

  • to make an image that looks quite like that one. So this is very like what's called the,

  • kind of watercolor model of images, where you know about where the boundaries are and

  • you just, sort of, roughly sketching the colors of the regions and it all looks fine to us,

  • because we sort of slaved the color boundaries to the actual--where the edges are. If you

  • reversed the colors of these it produce the reversed image because the covariance doesn't

  • care at all about the signs of things. If you look at the filters at your lens, the

  • mean units which are for sort of coloring in regions, learn these blurry filters--and

  • by taking some combination of a few dozen of those you can make more or less of what

  • other colors you like anywhere. So, very blur there-–smooth, blurry, and multicolored

  • and you can make roughly the right colors. The covariance units learn something completely

  • different. So, these are what the filters learned and you'll see that, those factors,

  • they learn high frequency black and white edges. And then a small number of them, turning

  • to low frequency color edges that are either red, green, or yellow blue and what's more

  • when you make it from a topographic map using a technique I'll describe on the next slide.

  • You get this color blob, this low frequency color blob in with the low frequency black

  • and white filters. And that's just what you see in a monkey's brain, pretty much. If you

  • go into a monkey's brain you'll see these high frequency filters whose orientation changes

  • smoothly as you go through the cortex tangentially, and you'll see these low frequency color blobs.

  • Most neuroscientists thought that at least must be innate. What this is saying is, "Nope.

  • Just the structure of images is, and the idea of forming a topographic map, is enough to

  • get this." That doesn't mean it's not innate, it just means it doesn't need to be. So the

  • way we get the topographic map is by this global connectivity from the pixels to the

  • factors. So the factors really are learning local filters. And the local filters start

  • off colored and gradually learn to be exactly black and white. Then there's local connectivity

  • between the factors in the hidden units. So one of these hidden units will connect to

  • a little square of factors and that induces a topography here and the energy function

  • is such that when you turn off one of these hidden units to say smoothness no longer applies,

  • you pay a penalty. And you derive to just pay the penalty once. And so two factors is

  • going to come on at the same time, it's best to connect them to the same hidden unit so

  • you only pay the penalty once. And so that will cause similar factors to go to similar

  • places in here when we get a topographic map. For people who know about modeling images,

  • as far as I know, nobody has yet produced a good model of patches of color images. That

  • is the genres of model that generates stuff that looks like the real data. So, here's

  • a model that was learned on 16x16 color images from the Berkeley database and here's these

  • generated from the model. And they look pretty similar. Now, it's a partly a trick, the color

  • balance here is like the color balance and it makes you think they are similar. But,

  • it's partly real. I mean, most of these are smooth patches of roughly uniform color as

  • are most of these. These are few more of these as smooth than those. But you also get these

  • things where you get fairly sharp edges, so you get smoothness than a sharp edge than

  • more smoothness, like you do in the real data. You even get things like corners here. We're

  • not quite there yet but this is the best model there is in patches of color images. And it's

  • because it's modeling both the covariance and the means, so it's capable of saying,

  • "What's the same as what?" As well as, "What the intensities are?" You can apply it for

  • doing recognition. So this is a difficult object recognition to ask where this 80 million

  • unlabelled training images; not only of these classes but of thousands and thousand of classes.

  • They were collected by people at MIT. It's called the Tiny Images database. They're 32x32

  • color images. But it's surprising what you can see in a 32x32 color image. And since

  • the biggest model we're going to use has about a hundred million connections, that's about

  • 0.1 of a cubic in ratio of cortex in terms of the number of parameters, and so we have

  • to somehow give our computer model some way of keeping up with the brain which has a lot

  • more hub, right? And so we do it by giving it a very small retina. We say, "Suppose the

  • input was only 32x32, maybe we can actually do something reasonable there." So as you'll

  • see there are a lot of variations. If you look birds, that's a close up of an ostrich,

  • this is a much more typical picture of a bird. And it's hard to tell the difference within

  • these tiny categories. Particularly things like deer and horse. We deliberately chose

  • some very similar categories like truck and car, deer and horse. People are pretty good

  • at these. People won't make very many errors. That's partly because these were hand-labeled

  • by people, so. But even people make some of errors. We only have 50,000 training examples.

  • Five thousand of each class and ten thousand test examples, because we have to hand-label

  • them, but we have a lot of untrained--unlabelled data. So we can do all these pre-training

  • on lots of unlabelled data and then take out covariance units on our mean units and just

  • try doing multi-[INDISTINCT] on top of those, or maybe add another hidden layer and do it

  • on top of that. So, what Marc'Aurelio Ranzato actually did since he worked in Yann LeCun's

  • lab, he actually took smaller patches learned the model and then strode them across the

  • image and replicated them. So it's a sort of a semi-convolutional. And then took the

  • hidden units of all of these little patches and just concatenated them to make a great

  • big vector of 11,000 hidden units which are both the means on the covariances. And then

  • we're going to use that as our features and see how well we can do. And we're going to

  • compare it with various other methods. So the sort of first comparison, you just take

  • the pixels and do logistic ration on the pixels to slide on the tiny glasses. You get 36%

  • right. If you take GIST features which has developed by Torralba and the people at MIT,

  • which were meant to capture what's going on under the image quite well, but they're fairly

  • low dimensional, you get 54%. So they're much better than pixels. If you take a normal RBM

  • which has linear units with glass and noises input variables and then binary hidden units,

  • and then you use those binary hidden units to do castigation, you get 60%. If you use

  • one of these RBMs with both the units like these once for doing the means, and then these

  • units with the three range interaction for modeling covariances, you got 69%; as long

  • as you use a lot of these factors. And if you then learn an extra hidden layer of 8,000

  • units--so now it's just that times that is a hundred million, so there's an extra hundred

  • million connections you learn there. But that's fine because it's unsupervised then you just

  • learn it on lots of data. You get up to 72%. And that's the best result so far on this

  • database. One final thing, you can take this model that was develop for image patches and

  • the student that'll be doing framing recognition just took that code and applied it to log

  • spectrograms, which is sort of more close to what they would like to see, you're not

  • using all these mark up fool stuff, which is designed to throw away stuff you think

  • you don't need and get rid of lots of correlations. Instead you're going to take data that has

  • lots of correlations in but we got a model that can deal with that stuff now. And the

  • first thing George tried on February the 20th, which was four layers of a thousand hidden

  • units on top of this, he got 22.7 correct--percent correct; which was the record for phoneme

  • recognition on the TIMIT database where your not trying to do a model adapted to each speaker.

  • And then a week later when he did that to TIMIT and use more frames, he was down to

  • 21.6%. So this--all this stuff was designed to do vision. It wasn't designed to do phonemes.

  • And if we treat phoneme recognition, it's just a vision problem on a lot of spectral

  • ground. We can wipe out the speech class, at least on small vocabulary. Another student

  • is now, at Microsoft, is seeing if this will work on big recovery as well.

  • >> [INDISTINCT] >> HINTONS: Yes. Yes, right.

  • >> We can give them new better tools. >> HINTONS: We can give them new and better

  • tools. So here's phoneme recognition over the years. Backprop from the 80's got 26.1

  • percent correct. Over the next 20 years or so, they got that down to 24.4 percent, using

  • methods that weren't learning-inspired so we'll call them artificial. We then got down

  • to 21.6 percent; an estimate of human performance is about 15 percent. I don't know much about

  • how they did this estimate, I'm afraid. But we're about--we're nearly a third of the way

  • from artificial TIMIT. And so we need two more ideas and we're there. Okay, I'm done.

  • I'm finished. >> Questions?

  • >> HINTONS: Yes? >> You mentioned YouTube recently announced

  • that the [INDISTINCT] have broken the world record on the end list of data sets of phoneme

  • recognition by simply using a seven layered feed forward network trained with backprop,

  • but doing it on a GPU with lots and lots of cycles.

  • >> HINTONS: Yes, he did indeed announce that. What he didn't announce was--he's got a spectacular

  • result. He gets timed to 35 errors. What he didn't announce was there's two tricks involved.

  • One trick is to use a big net with lot of layers in a GPU board. That trick by itself

  • wouldn't give you 35 errors. There's a second trick which was sort of pioneered by people

  • at Microsoft in fact, which is to put a lots of work into producing distortions of the

  • data so you have lots and lots of labeled data. So you take a labeled image of a two

  • and you distort it in clever ways and make it still look like a two but be translated

  • so people can then get down to about 40 errors. >> I think they patented that already.

  • >> HINTONS: Good. So Dick's already patented that. So you get down to--you can get down

  • to by 40 errors by doing these distortions. What he did was even better distortions, or

  • more of them, and a much bigger net on a GPU and he got from a 40 to 35, which is impressive

  • because it is hard to make any progress there. But it won't work unless you have a lot of

  • labeled data. And what's--the disguised thing is the work went into--if you look in the

  • paper, it's always straightforward, its just backprop, except when you get to the section

  • of how they generated all those sector labeled data where there's very careful things, like

  • if it's a one or a seven they'd only rotate it a certain number of degrees but if it's

  • something else they rotate it in more degrees. I'm actually the referee for this paper but

  • I don't mind him knowing. I think it's a very important work. But he should emphasize that

  • they have to have labeled data to do that, and they have to put work into distortions.

  • So for me the lesson of that paper is when we small computers, you should put your effort

  • into things like weight constraints so you don't have too many parameters because you

  • only got a small computer. As computer gets bigger and faster, you can transfer your effort

  • from, instead of tying the weights together, like Yann was doing in the early days, put

  • your effort into generating more distortions so you can inject your prior knowledge in

  • the form of distortions and that's much less complication-efficient over the big computers,

  • it's fine and it's more flexible. So I think that's the lesson of that paper.

  • >> I shouldn't even need to ask you a question, you answered it. Thank you.

  • >> HINTON: Any other long question? >> It seems like you've invented some kind

  • of a cortex here that house you expect the property that if it does vision it'll do sound.

  • >> HINTONS: Yes. >> What other problems you going to apply

  • it to? >> HINTONS: Maybe it'd be quicker to say the

  • problems we're not going apply. >> Okay.

  • >> HINTONS: I can't think of any. I mean--okay, let me say what the main limitation of this

  • is for vision. We got at least 10 billion neurons for doing visual things; or at least

  • a billion anyway, probably, 10 billion. And even if we got that many neurons and about

  • 10 to the 13 connections for doing vision, we still have a retina that's got a very small

  • phoneme the size of my thumb there at arms length. And so we still take almost everything

  • and don't look at it. I mean, the essence of vision is not to look at almost everything

  • intelligently; and that's why you got all this funny illusions where you don't see things.

  • We have to do that in these models. These models are completely crazy. And all of computer

  • visions are completely crazy, almost all of it. Because they take a uniform resolution

  • image, and quite a big one like a thousand by thousand, and they try and deal with it

  • all at once with filters all over the image. And if they going to do a selection, they

  • either do it by running off their face to get to everywhere, with no intelligence, or

  • they do sort of interest point detection at a very low level to decide what to attend

  • to. What we do is we fixate somewhere. Then on the basis of what our retina gives us,

  • with these big pixels around the edges and small pixels in the middle, we sort of decide

  • what we seeing and where to look next and by the second or third fixation we've fixating

  • very intelligently and the essence of it is that vision is sampling, it's not processing

  • everything; and that's completely missing from what I said. Now in order to do that,

  • you have to be able to do take what you saw and where you saw it and combined them and

  • that will multiply. So this module, it can do multiply. It's very good in combining what's

  • and where's, to integrate information at a time. And that's one of the things, we're

  • working on that. But that's probably the biggest thing missing. But that is an example of having

  • a module is quite good but now it's never good enough, so you have to put it together

  • over time and use it many times. And that's what sequential reasoning in all this stuff

  • are. So basically, as soon people become sequential we're not modeling that at all. We're modeling

  • what you can do in hundred milliseconds. And so that's what's missing. But I believe that

  • to model that sequential stuff we need to understand what is the sequence of, is the

  • sequence of these very powerful operations. And we're in a better shape now to try and

  • model sequential AI, than we were if we didn't know what a primitive operation is. So this

  • sort of primitive operation was just deciding whether two symbols are the same. We're going

  • to be out of luck for understanding how people do sequential stuff. Yeah.

  • >> This is a [INDISTINCT] question as he said he wanted to do everything if it connects.

  • Are you going to do [INDISTINCT] logic like there exists a God and every girl has a boy

  • she loves? >> HINTON: Hang on, I'm still processing that.

  • Right. Right, I'm making the point that people find "quantifies" quite difficult.

  • >> Oh, yeah. If you [INDISTINCT] quantifiers... >> HINTON: I would love to do that. I have

  • not got a clue how to do it. And you will notice that in old-fashioned AI that you used

  • to point out to [INDISTINCT] people, then you can't do quantifiers, so forget it. Nowadays,

  • when they all do graphical models, they didn't mention that anymore because the graphical

  • models have difficulty of it too. Some people has got [INDISTINCT] some people do. Right.

  • Yeah, some people do. But most of the graphical models of, like, five years ago, they do quantifiers

  • either. And so, a pretty good division line would be what you can do without having to

  • deal with really sophisticated problems like that. I would love to know how we deal with

  • that, but I don't. >> Thank you.

  • >> HINTON: So, yeah, I'm going to give up on that right now.

  • >> So if you had 80 million labeled images and no extra unlabeled ones, would you do

  • your pre-training... >> HINTON: Yes. Yes

  • >> ...and then fine tuning to make us better? >> HINTON: In TIMIT, that's what we have.

  • In TIMIT, all the examples we have labels. It stirs a big wind to do the pre-training.

  • >> But you didn't sneak this result I'm just hearing about? It seems to suggest...

  • >> HINTON: Well, the audience switched it but I haven't tried with all these distortions

  • during pre-training. Now, I've assumed student called [INDISTINCT] who just produced a thesis.

  • Well, he tries things like that. He tries distortions in earnest and he uses special

  • distortions of his own. And the fact is distortions helped a lot. But if you do pre-training,

  • that helps some more too. And [INDISTINCT] results, yes, [INDISTINCT] results, suggest

  • that pre-training will get you through different part of the space even if you have all these

  • labeled data. So clearly, one thing that needs to be done is to try the pre-training and

  • combine with these labels. You don't have to have the pre-training, but I bet you, it

  • still helps. And I bet you, it's more efficient too. It's faster because the pre-training

  • is rather pretty fast, you always have to learn a very good model. You got lots of its

  • features. And starting from there, I think, you'll do better than he does just started

  • from random, and faster. That's just a prediction. You might even get done to 34 out of this.

  • The problem with [INDISTINCT] you can't get significance. TIMIT is really nice that way.

  • They designed it well, so you get higher rates. So you can see differences.

  • >> On the time series aspect, did you see anything that would you get inferences or

  • alterations that are beyond the size of the time window you're using?

  • >> HINTON: Sorry, I didn't understand the question. We have a limited time. We don't...

  • >> You have limited time, after training is there anything of a model that picks up...

  • >> HINTON: Nothing. >> Nothing.

  • >> HINTON: Nothing. It cannot deal with--it can't model host...

  • >> It has an internal state. It has an internal state.

  • >> HINTON: Right. But if sort of what happened 15 times steps ago really tells you what should

  • happen now, and it only tells you what you should happen now. It doesn't tell you what

  • should happen in TIMIT 14 times steps. It just contains information across 15 times

  • steps without having a signature of smaller time scales. You can't pick up on that.

  • >> Okay. >> HINTON: Because it's not got a hidden forward-backward

  • algorithm. A forward-backward algorithm potentially could pick up a lot of load, actually can't.

  • >> So this one wouldn't pick up on things like object permanence or all rules behind

  • the box and comes out of the other side and they're not going to be able to...

  • >> HINTON: Not over a long time scale, no, no. Unless you say that there's a memory involved

  • when you go back to a previous--it gets more complicated, right? Now, it is true that when

  • you build the multilevel one, which you can do with the three interconnections as well

  • as with the three-way connections, at every level you're getting a bigger time span because

  • your time window, it's going further back into the past with each level. So you get

  • a bit high, but that's just sort of linear. >> Can you say--do you have any rules of thumb

  • of how much unlabeled data you need to train each of the different levels and how it would

  • change, like, is it just linear with the number of rates or as you go up levels the things

  • changed? >> HINTON: I have one sort of important thing

  • to say about that, which is that if you're modeling high-dimensional data and you're

  • trying to build an unsupervised model of the data, you need many less trainings on [INDISTINCT]

  • than you would have thought if you use the discriminative learning. When you're doing

  • discriminative learning, there's typically a very few bits per training case to constrain

  • the parameters. You're going to constrain--you got many new parameters for a training case

  • is the number of bits it takes to specify the answer, not the number it takes to specify

  • the input. So within this, you get 3.3 bits per case. If you're modeling the image, the

  • number of bits per case is the number of bits it takes to specify to image which is about

  • a hundred bits. So you need far fewer cases per parameter. In other words what I'm saying

  • is you're modeling much which are things, and so each case is giving you much more information.

  • So actually, we can typically model many more parameters than we have training cases. And

  • discriminative people aren't used to that. Many less parameters than we have pixels and

  • many more than training cases. And in fact, he used about two million cases for doing

  • the image stuff, and it wasn't enough, it was over fitting. He should have used more.

  • But he was fitting 100 million parameters. But the--basically, the only rule of thumb

  • is many less parameters and the number of total number of pixels in your training data,

  • but you can typically use many more parameters in the number of training cases. And you can't

  • do that with normal discriminative learning. Now, if you do do that, when you start discriminative

  • training, it quickly improves things and then very quickly over fits. So you have to stop

  • it early. Okay. >> Okay?

  • >> HINTON: Thanks. >> Let's thank the speaker again.

  • >> Thank you.

>>

字幕と単語

ワンタップで英和辞典検索 単語をクリックすると、意味が表示されます

B1 中級

ディープラーニングの最近の動向 (Recent Developments in Deep Learning)

  • 174 14
    Howard Liu に公開 2021 年 01 月 14 日
動画の中の単語