字幕表 動画を再生する 英語字幕をプリント >> So it's my pleasure to introduce to you Geoff Hinton, who is a pioneer in machine learning and neural nets, and more recently [INDISTINCT] architectures. And then, I think that's going to be topic of today. So take it over. >> HINTON: Okay. So I gave it to Okeer a couple of years ago. And the first 10 minutes or so will be an overview of what I said there, and then I'll talk about the new stuff. The new stuff consists of a better learning module. It allows you to learn better in all sorts of different things, like, learning how images transform, learning how people walk, and learning object recognition. So the basic learning module consists of some variables that represent things like pixels, and these will be binary variables for that. Some variables that represent--these are latent variables, they're also going to be binary. And there's a bipartite connectivity, so these guys are connected to each other. And that makes it very easy if I give you the states of the visible variables to infer the states of the hidden variables. They're all independent given the visible variables because it's a non-directed graph. And the input procedure just says, the probability of turning on hidden unit "hj" given this visible vector "v" is the logistic function of the total what he gets from his audience, so very simple for the hidden variables. Given the hidden variables, we can also infer the visible variables very simply. And if we want some--if we put some weights on the connections and we want to know what this model believes, we can just go back and then forward inferring all the hidden variables in parallel than all the visible ones. Do that for a long a time, and then you'll see examples of the kinds of things it likes to believe. And the end of learning is going to be to get it to like to believe the kinds of things that actually happen. So this thing is governed by an energy function that has given the weights on the connections. The energy of a visible plus a hidden vector is the sum overall connections of the weight if both the visible and hidden units are active. So I'm going to pick some of the features that are active, you adding the weight, and if it's a big positive weight, that's low energy, which is good. So, it's a happy network. This has nice derivatives. If you differentiate it with respect to the weights, you get this product of the visible and hidden activity. And so, that the derivative is going to show up a lot in the learning because that derivative is how you change the energy of a combined configuration of visible and hidden units. The probability of a combined configuration, given the energy function, is E to the minus the energy of that combined configuration normalized by the partition function. And if you want to know the probability of a particular visible vector, you have to sum all the hidden vectors that might go with it and that's the probability of visible vector. If you want to change the weights to make this probability higher, you always need to lower the energies of combinations of visible vector on the hidden vector that would like to go with it and raise the energies of all other combinations, so you decrease the computation. The correct maximum likelihood learning rule that is if I want to change the weights so as to increase the log probability, that this network would generate the vector "v" when I let it just sort of fantasize the things it like to believe in is a nice simple form. It's just the difference of two correlations. So even though it depends on all the other weights, it shows up this is difference of correlations. And what you do is you take your data, you activate the hidden units, that's to classify the units, and then we construct, activate, we construct activate. So this is a mark of chain. You run it for a long time, so you forgot where you started. And then you measure the correlation there, start with the correlation here. And what you're really doing is saying, "By changing the weights in proportion to that, I'm lowering the energy of this visible vector with whatever hidden vector it chose. By doing the opposite here, I'm raising the energy, the things I fantasize." And so, what I'm trying to do is believe in the data and not believe in what the model believes in. Eventually, this correlation will be the same as that one. In most case, nothing will happen because it will believe in the data. In terms that you can get a much quicker learning algorithm where you can just go on and [INDISTINCT] again, and you take this difference of correlations. Justifying that is hard but the main justification is it works and it's quick. The reason this module is interesting, the main reason it's interesting is you can stack them up. That is for accompanied reason you're not going to go into it, it works very well to train the module then take that activities of the feature detectors, treat them as so they were data, and train another module on top of that. So the first module is trying to model what's going on in the pixels by using these feature detectors. And the feature detectors would tend to be highly correlated. The second model is trying to model a correlation among feature detectors. And you can guarantee that if you do that right, every time you go up a level, you get a better model of the data. Actually, you can guarantee that the first time you go up a level. For further levels, all you can guarantee is that there's a bound on how good your model of the data is. And every time we add another level, that bound improves if we had it right. Having got this guarantee that something good is happening as we add more levels, we then violate all the conditions of mathematics and just add more levels in sort of [INDISTINCT] way because we know good things are going to happen and then we justify by the fact that good things do happen. This allows us to learn many lesser feature detectors entirely unsupervised just to model instruction of the data. Once we've done that, you can't get that accepted in a machine learning conference because you have to do discrimination to be accepted in a machine learning conference. So once you've done that, you add some decision units to the top and you learn the connections discriminatively between the top-layer features and the decision units, and then if you want you can go back and fine-tune all of the connections using backpropagation. That overcomes the limit of backpropagation which is there's not much information in the label and it can only learn on label data. These things can learn on large amounts of unlabeled data. After they've learned, they you add these units at the top and backpropagate from this small amount of label data, and that's not designing the feature detectors anymore. As you probably know at Google, designing feature detectors is the art of things and you'd like to design feature detectors based on what's in the data, not based on having to produce labeled data. So the edge of backpropagation was design your feature detectors so you're good at getting the right answer. The idea here is design your feature detectors to be good at modeling whatever is going on in the data. Once you've done that, just have a so slightly fine-tune and so you better get right answer. But don't try and use the answer to design feature detectors. And Yoshua Bengio's lab has done lots of work showing that this gives you better minima than just doing backpropagation. And what's more minima in completing different part of the space? So just to summarize this section, I think this is the most important slide in the talk because it says, "What's wrong with million machine learning up to a few years ago." What people in machine learning would try to do is learn the mapping from an image to a label. And now, it would be a fine thing to do if you felt that images and labels are rows in the following way. The stuff and it gives rise to images and then the images give rise to the labels. Given the image, the labels don't depend on the stuff. But you don't really believe that. You only believe that if a label is something like the parity of the pixels in the image. What you really believe is the stuff that gives rise to images and then the labels that goes with images because of the stuff not because of the image. So it's a cow in a field and you say cow. Now, if I just say cow to you, you don't know whether the cow is brown or black, or upright or dead, or far way. If I show an image of the cow, you know all those things. So this is a very high bandwidth path, this is a very low bandwidth path. On the right way to associate labels with images is to first learn to invert this high bandwidth path. And we can currently do that because vision works basically. The first store you look at then, you see things. And it's not like it might be a cow, it might be an elephant, it might be electric theater. Basically, you get it right nearly all the time. And so we can invert that pathway. Having learned to do that, we can then learn what things are called. But you get the concept of a cow not from the name, but from seeing what's going on in the world. And that's what we're doing and then later as I say from the label. Now, I need to do one slight modification to the basic module which is I had binary units as the observables. Now, we want to have linear units with Gaussian noise. So we just change the energy function of it. And the energy now says, "I got a kind of parabolic containment here." Each of these linear visible units has a bias which is like its mean. And it would like to sit here and moving away from that [INDISTINCT] energy. The parabola is the negative log of the Gaussian [INDISTINCT]. And then the input that comes from the hidden units, this is just vi, hj, wij, but Vs have to be scaled by the standard deviation of the Gaussian there. If I ask--if I differentiate that with respect to a visible activity, then what I get is hj, wij divided by the sigma I. And that's like an energy gradient. And what the visible unit does when you reconstruct is it tries to compromise between wanting to sit around here and wanting to satisfy this energy gradient, so it goes to the place where this two gradients [INDISTINCT] opposite and you have--that's the most likely value and then you [INDISTINCT] there. So with that small modification we can now deal with real value data with binary latent variables and we have an efficient learning algorithm that's an approximation of [INDISTINCT]. And so we can apply it to something. So it's a nice speech recognition task that's been well organized by the speech people where there's an old database called TIMIT, it's got a very well-defined task for phone recognition where what you have to do is you're given the short window speech, you have to predict the distribution, the probability for the central frame of the various different phones. Actually, each phone is modeled by 3-state HMMs, sort of beginning middle and end, so you have to predict for each frame is it the beginning middle or end of each with the possible phones, there's a 183 of those things. If you give it a good distribution there to sort of focus on the right thing then all the post-processing will give you back where the phoning bandwidth should be and what your phone arrow radius, and that's all very standard. Some people use tri-phone models. We're using bi-phone models which aren't quite as powerful. So now we can test high goodwill by taking 11 frame of speech. It's 10 milliseconds per frame but each frame is looking at like 25 milliseconds of speech and predicting the phone at the middle frame. We use the standard speech representation which is mel-cepstral coefficients. There's 30 of those, and there are differences and differences and difference, differences; and we feed them in to one of these deep nets so. So here's your input, 11 frames and 39 coefficients. And then--I was away when the student did this and he actually believed what I said. So he thought adding lots and lots of hidden units was a good idea. I've started it too. But he added lots of hidden units all unsupervised, so all this green connections are learned without any use of the labels. He used to bottleneck there, so the number of Reg connections will be relatively small. These are not--these have to be learned using discriminative information. And now you're back propagating the correct answers through this whole net for about a day on a GPU board or a month on a core, and it does very well. That is the best phone error rate we got was 23%. But the important thing is whatever configuration you use, how many hidden layers as long as they are plenty and whatever widths and whether you use this bottleneck or not, it gets between 23% and 24%. So it's very robust to the exact details of how many layers and how wide they are. On the best previous result on TIMIT for things that didn't use speaker adaptation is 24.4% and that was averaging together on lots of models, so this is good. >> So each of these layers that's four million weights? >> HINTON: Yup, four million weights. So we're only training one, two, three, one, two, three, we're training, you know, about 20 million weights. Twenty million weights is about 2% of a cubic millimeter of cortex. I think so, this is a tiny brain. The last, probably, all you need for [INDISTINCT] recognition. >> Why did they start with the differences and double differences of the MFCCs that you're going into a thing that could learn to do that itself if they wanted to? >> HINTON: That's a very good question 'cause you are sitting at the end. It's an extremely good question because the reason I put the differences and double differences is so they can model a data with a diagonal co-variance metric--diagonal co-variance model--and you can't model the fact that overtime two things turn to be very much the same where modeling co-variance is, unless you actually put the differences into the data and you model the differences directly. So it allows you to use a model that conquered with co-variances. Later on we're going to show a model that conquered with co-variances and then we are going to do what the client always said you should do, which is throw away the mel-cepstral representation and use the better representation in speech. >> I said that? >> HINTON: Yes, you said that to me the last time [INDISTINCT]. >> Smart guy. >> HINTON: Okay, so the new idea is to use a better kind of module. This module already works pretty well, right? You know, it does well at forming recognition, it does well in all sort of things. It can't model multiplicative interactions very well. It can model anything with enough training data, but its not happy modeling multiplies. It multiplies all over the place. I'll show you a bunch of place where you need multiplies. Here's the sort of main example of why you need multiplies. Supposed I want to, from a high level of description of an object the name of the shape and it's pose, size, position orientation. So, first of all, I want to generate the parts of an object and I want them to be related correctly to each other. I could use very accurately top down model that says none of this square and none pose gram, because I generate each piece in exactly the right position; that would require high bandwidth. Or I could be sloppy and I could say, "I need to generate this side, not sort of a representation and distribution of where this side might be. And I'll generate corners and other sides and they're all a bit sloppy. And if I picked one thing from each distribution, it would make a nice square. But I could also top-down specify how these things should be pieced together. In effect, I can specify a macro form of field; this is what goes with what? And then I can clean this up knowing these distributions and pick a square like that. Of course, I might sometimes pick a square that has slightly different orientation or slightly different size, but it'll be a nice clean square because I know how they go together. And so that's a much more powerful kind of generation model, and that's what we want to learn to do, and so we are going to need hidden units up here to specify interactions between visible units here, as opposed to just specifying input of visible units. There's an analogy for this, which is, if I'm an officer and there's a bunch of soldiers and I want them to stand in the square, I could get out my GPS and I can say, "Soldier number one, stand at this GPS coordinates. And soldier number two, stand at these GPS coordinates." Now, if I use enough digits, I'll get a nice neat rectangle, or I could say, "Soldier number one, stand roughly around here. And then soldier number two, hold your arm out and stand this distance from soldier number one." And that's a much better way to get a neat rectangle. It will cause far less communication. So what you're doing is you're downloading roughly where people should stand and then how they should relate to each other. We have to specify the relations not just where they should be. And that's what we'd like in a powerful [INDISTINCT] model. So, we're going to aim to get units in one layer to say how units in the layer below should latterly interact when you generate it. It's going to turn out you don't need to worry about these lateral interactions when you recognize it. When you generate, you do. To do that, we're going to need things called third-order Boltzmann machines, which has three lane tractions. So Terry Sejnowski, he pointed out a long time ago that we have an energy function like this where this was V and this was H, but these are just binary variables. And we could probably well write down an energy function like this with three things in tract, then we have a three layer weight. And if you think about these three things now, K, the state of K is acting like a switch. When K is on, you effectively have this weight between I and J. When K is off, this weight disappears. And it happens every which way because it's symmetric. So using an energy function like this, we can allow one thing to specify how two other things should interact. So each hidden unit can specify a whole mark of random field over the pixels if you want. But that sort of begins to make you worry because a mark of random field has a lot of parameters in it. And if you start counting in the sits here, if you have any of these and any of those and none of those, you get enqueued to these parameters which is rather a lot. If you're willing to use enqueued parameters, you can now make networks to look like this. Suppose I have two images and I want to model how images transform over time, unless suppose I'm just moving random dots around, have a pan of random dots and I translate it. Well, if I see that dot and I see that dot, that's some evidence for a particular translation. And so if I put a big positive weight there, this triangle is meant to interrupt that big three-way weight. Then when this and this around, they'll say it's very good to have this guy up. It would have been nice at low energy state. If I also see this pair of dots, I'll get more routes, though, and this guy should be--and I will turn this guy on. If however this pixel went to here, I'll go for this guy. And if this pixel also went to there but this guy--so these guys are going to represent coherent translations of the image, and it's going to be able to use these three-way weights to take two images and extract two units that represent the coherent translation. It'll also be able to take the pre-image and the translation, and compute with pixel should be on here. Now what we're going to do is take that basic model and we're going to factorize it. We're going to say, "I've got these three-way weights and I've got too many of them." So, I'm going to represent each three-way weight as the product of three two-way things. I'm going to introduce these factors and each factor is going to have these many parameters which is just pro-factor, is just a linear number of parameters. If I have about N factors, I end up with only N square of these weights. And if you think about how pixels transform in a new image, they don't do random permutations. It's not that this pixel goes on that one, goes here. Pixels do sort of consistent things, so I don't really need enqueued parameters because I'm just trying to model these fairly consistent transformations, which is a limited number, and I should be able, too, in many less parameters. And this is the way to do it. So, that's going to be our new energy function given the bias terms. One way of thinking about how modeling a weight is I want these tensor of three-way weights. If I take an IF product and two vectors like this, I'll get a matrix that has rank one. So I get a three-way product. I'll get a tensor that has rank one. And if I now add up a bunch of tensors like that so each factor now, each F, specifies a rank one tensor, by adding up a bunch of them, I can model any tensor I like if I use N square factors. If I use any N factors, I can model most regular tensors but I can't model arbitrary permutations, and that's what we want. If you ask how does inference works now, inference is still very simple in this model. So here's a factor. Here's the weights connecting it to, say, the pre-image. Here's the weights connecting it to the post-image. Here's the weights connecting it to the hidden units. And to do inference what I do is this. Supposed I only have that one factor. I would multiply the pixels by these weights; add all that up so I get a sum of this vertex. I do the same here; I get a sum of this vertex. Then I multiply these two sums together to get a message going in the center of the hidden units. And as that message goes to the hidden unit, I multiply it by the weight on my connection. And so what the hidden unit will see is this weight, turns the product of these two sums, and that is the derivative of the energy with respect to the state of this hidden unit, which is what it needs to know to decide whether to be on or off, it wants to go into what other state below the image. And all the hidden units remain independent even though I've got these multipliers now. So this is much better than putting in another stochastic binary unit here. If I put a stochastic binary unit in here, the hidden units would cease to be independent and inference will get tough. But this way, whether a deterministic factor that's taking a part of these two sums, inference remains easy. The learning also remains easy. So this is the message that goes from factor F to hidden unit H, and that message is the product that we got of these two lower vertices; the product of the sums, the compute on the pre-image and the post-image. And the way you learn the weight on the connection from factor F to hidden unit H is by changing the weight so as to lower the energy when you're looking at data, and raise the energy when you're constructing these from the model or just reconstructing things from the hidden units you got from data. And those energy groups, they just look like this. They're just the product of the state of the hidden unit and the message that goes to it when you're looking at data and the state of the hidden unit and the message that goes through it when you're looking at samples from the model or reconstructions. So it's still a nice pair-wise learning rule. So everything is pair-wise still, so you might fit into the brain. Now, if we look what one of these factors does when I show random dot patterns to translate, then we can look at the weights connecting it to the pre-image, and that's a pattern of weights where white is a big positive weight, black is a big negative weight because that would have a learned a great in connecting it to the pre-image and this will have learned a great in connecting it to the post-image. With a hundred factors, I'll show you what Roland learned. So, those are the hundred the hundred factors connecting--these are the receptive fields at the factors in the pre-image. And remember it's looking at translating dots, and these are the factors in the post-image. And you see, it's basically learned the freer basis and it's learned to translate things by about 90 degrees. And that's a very good way of handling translation. Mathematicians say things like, "The freer basis is a natural basis for modeling translation." I don't really know what that means, but just learn the freer basis on that. And if you get rotations, it learned the different basis. So this is the basis that learns for rotations. You see it learns about yin and yang here. Oops [INDISTINCT]. Okay, that's the basis for rotations. One other thing you could do is train it just on single dot pans and translating in a coherent way and then test it on two overlaid dot patterns but they're translating different directions. It's never seen that before. It's only been trained on coherent motion where we're going to test it on what's called transparent motion. In order to see what it thinks, when we train the uncivilized, there's no labels anywhere, we never tell it what the notions are, we need some way to seeing what it's thinking, so we add a second hidden layer that looks at the hidden units representing transformations and it's fairly sparse. So the units on that second hidden layer will be tuned to particular directions of motion. And then to see what it's thinking, we take the directions those units like weighted by how active those units are and I will tell you what directions it thinks it's seeing. Now when you show it transparent motion and you look at those units in the second hidden layer, if the two motions are within about 30 degrees, it sees a single motion of the average direction. If they're beyond about 30 degrees, it sees two different motions and once more they're repelled from each other. That's exactly when I was with people, and so this is exactly how the brain works. Okay. There's going to be a lot of that kind of reasoning in this talk. I'm going to on to time series models now. So, we'd like to model not just static images, for example, we like to model video. To be [INDISTINCT] we're going to try something a bit simpler. When people do time series models, you would nearly always like to have a distributed non-linear representation, but that's hard to learn. So people tend to do dumb things like Hidden Mark up Models or Linear Dynamical Systems which either give up on the distributed or on the non-linear, but are easy to doing inference. What we're going to come up with is something that has the distributed and the non-linear and is easy to do inference, but the linear algorithm isn't quite right but it's good enough. It's just an approximation to make some [INDISTINCT]. And the inference also is ignoring the future and just basing things on the past. So, here's a basic module, and this is with just two-way interactions. This is the Restricted Boltzmann Machine with visible units and hidden units. Here are the previous visible frames. These are all going to be linear units. And so, these blue connections are conditioning the current visible values on previous observed values in a linear way. So, it's called an autoregressive model. The hidden units here are going to be binary hidden units; they're also conditioned on previous visible frames, and learning is easy in this model. What you do is you take your observed data, and then given the current visible frame and given the previous visible frames, you got import to the hidden units, they're all independent given the data, so you can separately decide what states they should be in. Once you fixed states for them, you now reconstruct the current frame using the input you're getting from previous frames and using the top you got in from the hidden units. After we construct, you then activate the hidden units again. When you say the difference in the power statistics with data here and the reconstructions here to learn these weights and you take the difference on activities of these guys with data with reconstructions to get signal that you can used to learn these weights or these weights. So learning is straightforward and it just depends on differences, and you can learn a model like this. After you've learned it, you can generate from the model by taking some previous frames. These inputs, the conditioning inputs, in effect, fixed the biases of these to depend on the previous frames. So, these are the dynamic biases, and with these biases fixed, you just get backwards and forwards for awhile and then pick a frame there, and that's your next frame you regenerated, then you keep going. So, we can generate from the model once it learns so we can see what it believes. >> You always go back two steps in time or is that just an example? >> HINTON: Sorry. >> Oh, you were just going back only two steps in time? >> HINTON: No, we're going to get back more sets in time. >> Okay, and you let... >> HINTON: I just got lazy with the PowerPoint. Now, one direction we could go from here is to higher level models. That is, having learned this model where these hidden units are all independent given the data, we could say--well, what I've done is I've turned the visible frames into the hidden frames map. And it turns out you can get a better model if you take these hidden frames, a model what's going here, and now you put in conditioning connections between the hidden frames and more hidden units that don't have conditioning here that don't interact with other hidden unit. [INDISTINCT] in this model. Then you can prove that if you do this right, then you'll get a better model of the original sequences or your improver band on the model of the original sequences. So you can [INDISTINCT] lots of layers like that. And when you have more layers, it generates better. But I'm going to go in a different direction. I'm going to show you how to do it with three-way connections. And we're going to apply it to motion-capture data, so you put reflective markers on the joints, you have lots of infrared cameras, you figure out where the joints are in space. You know the shape of the body so you go backwards through that to figure out the joint angles and then the frame of data is going to consist of 50 numbers, about 50 numbers which are joint angles and the translations and rotations of the base of the spine. Okay. So, imagine we got--one of these mannequins you see in art shop windows, we got a pins stuck in the base of his spine and we can move him around and rotate him using this pin and we can also wiggle his legs and arms. Okay. And what we want him to do is as we move him around, we want him to wiggle his legs and arms so his foot appears to be stationary on the ground and he appears to be walking. And he'd better wiggle his leg just right as we translate his pelvis, otherwise his foot will appear to skid on the ground. And we're going to model him, we can do hierarchal model like I just showed you or we can a three-way model like this where we condition six earlier of frames, this is a current visible frame, here's basic bolts in the machine accept that it's neither one of these 3-way things where these are factors. And we have a 1-of-N style variable. So, we have data and we tell it the style when we're training it, so that's sort of semi-supervise. It learns to convert that 1-of-N representation in to a bunch of real value features and then it uses this real value features as one of the inputs to a factor. And what the factor are really doing is saying, these real value features are modulating the weight matrixes that use for conditioning and also this weight matrix that use in your parallel linear model. So, these are modulating an auto aggressive model. That's very different from switching between auto aggressive model it's much more powerful. Yeah? >> I missed with what this one event is...? >> HINTON: So, we're going to have data of someone walking in various different styles. >> Styles of walking. >> HINTON: The style of walking. Yeah. >> So you mean your earlier diagram when you can't beat history, it looked like there was nothing to keep track in the relative order of the earlier direct delta direct link because... >> HINTON: Yes. >> ... is there anything in the model that cares about that relative...? >> HINTON: Yeah. Yeah. The weights on the connections will tell you which frame it's coming from. Right. In the earlier model, there were two blue lines, they're different matrixes and they have different weights on. >> There's nothing from two steps to previews to ones step previews, right, it just skip all the way? >> HINTON: It just skipped all the way, right. You just... >> Will that continue to happen? >> HINTON: Yes. In other words there's direct connections from all six previous frames to the current frame for determining the current frame. >> Right. And then what links from the six frames of the fifth earliest report? >> HINTON: Well, there where when you were computing what the fifth frame was doing, right? >> Okay. >> HINTON: But when we're computing this frame we have direct connections from it. Okay. So, we're now going to train this model, it's relatively easy to train especially on the GPU board, and then we're going to generate from it, so we can see sort of what it learned and we can judge if it's doing well by whether the feet slip on the ground. All right. >> [INDISTINCT] >> HINTON: We'll get there. >> Sorry. >> HINTON: Here's a normal walk. Maybe, at least they're willing. Okay. So I was generating from the model--he's deciding which direction to turn in, and he's deciding, you know, he needs to make the outside leg go farther than the inside leg and so on. If we--we have one model but if we flip the style label to say, gangly teenager, he definitely looks awkward. Right. We've all been there. I think this is a computer science student. My main reason for thinking that is if you asked him do a graceful walk, it looks like this. And that's definitely C3PO. >> [INDISTINCT]. >> HINTON: Now, I think this was a student [INDISTINCT]--but he's very good. You can ask him to walk softly like a cat. We're asking to model at present, right? The model looks pretty much like the real data the real data obviously the feet are planted better but notice, he can slow down then speed up again. Auto aggressive models can't do things like that, Auto aggressive models have a biggest size in value, the size is bigger than one in which case they explode or a smaller one in which case they die and the way to keep them alive is by keep--you keep injecting random noise so that they stay alive and that's like making a horse walk by taking a dead horse and jiggling it, it's kind of--it's not good. Now, he doesn't have any model of the physics so, in order to do this kinds of stumbles, there had to be stumble similar to that in the data but when he stop in which he stumble he did when, he's entirely determining. We could make him do a sexy walk but you're probably not interested in that. >> I just order a chicken. >> HINTON: You want dinosaur the chicken? Where's dinosaur the chicken? >> And chicken, number five. >> At number five. >> HINTON: Oh, no, that's dinosaur and chicken. That's a blend. Maybe a switch. He's got quite a lot of foot [INDISTINCT] that's probably a blend. This is doing a sexy walk and then you flip the label to normal and then you flip it back to sexy. It's never seen any transitions but because all one model, it can do reasonable for transitions. >> So you have these hundred style variables, can you de-couple those form the one event style and just make up new styles by playing with those... >> HINTON: Yup. Yup. Now, you can also give it many more labels when you train, you can give it speed, stride length all sorts of things then you can control it very well, yeah. Okay. So, you can learn time series at least for 50 dimensional data and obviously what we all want to do is apply that to video but we haven't done that yet. Except for some very simple cases. The last thing I'm going to show is the most complicated use of these 3-way models. One way of thinking of it, so that it's similar to the previous uses, is that we take an image and we make two copies of it but they have to be same. And then we insist the weights that go from a factor of this copy are the same as the weights that go from the factor of this copy. So if I=J, WI=WJF. Inference is still easy in fact inference here will consist of--you take these pixels times these weights to get a weighted sum, and then you square it because this is going to be the same weighted sum. So, inference consist--take linear filter, square its output, and send it by these weights to the hidden units. That's exactly the model called the [INDISTINCT] energy model, is right, kind of linear filter. This is being proposed both by Vision people by Adelson and Bergen, a long time ago in the '80s, and by neuroscientists. So, neuroscientists had tried to take simple cells, my point vaguely about, and look at what polynomial they're output is of their input, and Yang Dan at Berkeley says it's between 1.7 and 2.3, and that's means two. So, this looks quite like models that were proposed for quite different reasons and it just drops out of taking a 3-way imaging model and factorizing it. The advantage we have is that we have a learning algorithm for all these weighs now, when we generative model. So now we can model covariances between pixels, and the reason that's good is–-well, here's one reason why it's good. Suppose I asked you to define a vertical edge. Most people will say, "Well, vertical edge is something that light on the side and dark on that side. Well no, maybe it's light on this side and dark on that side, but you know. Well, it could be light up here and dark down there, and dark up here and light down there." Okay. Or it could be texture edge. It's getting--oh, it might actually be a disparity edge. Well, the manner should be motion this side and no motion that side. That's a vertical edge too. So, a vertical edge is a big assortment of things, and what all those things have in common is vertical edge is something where you shouldn't do horizontal interpolation. Generally new image, horizontal interpolation works really well. A pixel is the average of its right and left neighbors, pretty accurately almost all the time. Occasionally it breakdowns, and the place it breaks down is where there is a vertical edge. So, a real abstract definition of vertical edge is breakdown of horizontal interpolation. And that's what our models are going to do. A hidden unit is going to be putting in interpolation, and it's actually going to turn-off sort of reverse logic, when that breaks down its going to turn-off, so one way of seeing it is this. If this hidden unit here is on, it puts in a weigh between pixel I and pixel J that's equal to this weight times this weight, times this weight. Okay. Since these–okay, that's good enough. So, these are controlling affectively the mark of random field between the pixels, so we can model covariances nicely. Because the hidden units are creating correlations between the visible units reconstruction is now more difficult. We could reconstruct one image given the other image, like we did with motion, but if you want to reconstruct them both and make them identical it gets to be harder. So, we have to use a different mathical Hybrid Monte Carlo. Essentially you start where the data was and let it wonder away from where it was but keeping both images the same. And I'm not going to go to Hybrid Monte Carlo, but it works just fine for doing the learning. And the Hybrid Monte Carlo is used just to get the reconstructions and the learning algorithm just the same as before. And what we're going to do is we're going to have some hidden units that are using these 3-way interactions to model covariances between pixels, and other hidden units are just modeling the means. And so we call-–for meaning covariance, we call this mcRBM. Here's an example of what happens after it's learned on black and white images. Here's an image patch. Here's its reconstruction of the image patch, if you don't have noise, which is very good, from the mean and covariance units. Here's the stochastic reconstruction which is also pretty good. But now we're going to do something funny, we're going to take the activations of the covariance units. The things that are modeling which pixels are the same with which other pixels and we're going to keep those. But we are going to take the activations of the mean units, so we're going to throw those away, and pretend that the means from the pixels look like this. Well, let's take this one first. We tell all the pixels have the same value, except these are which are much darker and it now tries to make that information about means fit in with these information about covariances which is of these guys should be the same but very different from these guys. And so, it comes up with the reconstruction that it looks like that. Where you see it's taken this dark stuff and blurred across this region here. If we just give it four dots like that, and the covariance matrix you've got from there, it'll blur those dots out to make an image that looks quite like that one. So this is very like what's called the, kind of watercolor model of images, where you know about where the boundaries are and you just, sort of, roughly sketching the colors of the regions and it all looks fine to us, because we sort of slaved the color boundaries to the actual--where the edges are. If you reversed the colors of these it produce the reversed image because the covariance doesn't care at all about the signs of things. If you look at the filters at your lens, the mean units which are for sort of coloring in regions, learn these blurry filters--and by taking some combination of a few dozen of those you can make more or less of what other colors you like anywhere. So, very blur there-–smooth, blurry, and multicolored and you can make roughly the right colors. The covariance units learn something completely different. So, these are what the filters learned and you'll see that, those factors, they learn high frequency black and white edges. And then a small number of them, turning to low frequency color edges that are either red, green, or yellow blue and what's more when you make it from a topographic map using a technique I'll describe on the next slide. You get this color blob, this low frequency color blob in with the low frequency black and white filters. And that's just what you see in a monkey's brain, pretty much. If you go into a monkey's brain you'll see these high frequency filters whose orientation changes smoothly as you go through the cortex tangentially, and you'll see these low frequency color blobs. Most neuroscientists thought that at least must be innate. What this is saying is, "Nope. Just the structure of images is, and the idea of forming a topographic map, is enough to get this." That doesn't mean it's not innate, it just means it doesn't need to be. So the way we get the topographic map is by this global connectivity from the pixels to the factors. So the factors really are learning local filters. And the local filters start off colored and gradually learn to be exactly black and white. Then there's local connectivity between the factors in the hidden units. So one of these hidden units will connect to a little square of factors and that induces a topography here and the energy function is such that when you turn off one of these hidden units to say smoothness no longer applies, you pay a penalty. And you derive to just pay the penalty once. And so two factors is going to come on at the same time, it's best to connect them to the same hidden unit so you only pay the penalty once. And so that will cause similar factors to go to similar places in here when we get a topographic map. For people who know about modeling images, as far as I know, nobody has yet produced a good model of patches of color images. That is the genres of model that generates stuff that looks like the real data. So, here's a model that was learned on 16x16 color images from the Berkeley database and here's these generated from the model. And they look pretty similar. Now, it's a partly a trick, the color balance here is like the color balance and it makes you think they are similar. But, it's partly real. I mean, most of these are smooth patches of roughly uniform color as are most of these. These are few more of these as smooth than those. But you also get these things where you get fairly sharp edges, so you get smoothness than a sharp edge than more smoothness, like you do in the real data. You even get things like corners here. We're not quite there yet but this is the best model there is in patches of color images. And it's because it's modeling both the covariance and the means, so it's capable of saying, "What's the same as what?" As well as, "What the intensities are?" You can apply it for doing recognition. So this is a difficult object recognition to ask where this 80 million unlabelled training images; not only of these classes but of thousands and thousand of classes. They were collected by people at MIT. It's called the Tiny Images database. They're 32x32 color images. But it's surprising what you can see in a 32x32 color image. And since the biggest model we're going to use has about a hundred million connections, that's about 0.1 of a cubic in ratio of cortex in terms of the number of parameters, and so we have to somehow give our computer model some way of keeping up with the brain which has a lot more hub, right? And so we do it by giving it a very small retina. We say, "Suppose the input was only 32x32, maybe we can actually do something reasonable there." So as you'll see there are a lot of variations. If you look birds, that's a close up of an ostrich, this is a much more typical picture of a bird. And it's hard to tell the difference within these tiny categories. Particularly things like deer and horse. We deliberately chose some very similar categories like truck and car, deer and horse. People are pretty good at these. People won't make very many errors. That's partly because these were hand-labeled by people, so. But even people make some of errors. We only have 50,000 training examples. Five thousand of each class and ten thousand test examples, because we have to hand-label them, but we have a lot of untrained--unlabelled data. So we can do all these pre-training on lots of unlabelled data and then take out covariance units on our mean units and just try doing multi-[INDISTINCT] on top of those, or maybe add another hidden layer and do it on top of that. So, what Marc'Aurelio Ranzato actually did since he worked in Yann LeCun's lab, he actually took smaller patches learned the model and then strode them across the image and replicated them. So it's a sort of a semi-convolutional. And then took the hidden units of all of these little patches and just concatenated them to make a great big vector of 11,000 hidden units which are both the means on the covariances. And then we're going to use that as our features and see how well we can do. And we're going to compare it with various other methods. So the sort of first comparison, you just take the pixels and do logistic ration on the pixels to slide on the tiny glasses. You get 36% right. If you take GIST features which has developed by Torralba and the people at MIT, which were meant to capture what's going on under the image quite well, but they're fairly low dimensional, you get 54%. So they're much better than pixels. If you take a normal RBM which has linear units with glass and noises input variables and then binary hidden units, and then you use those binary hidden units to do castigation, you get 60%. If you use one of these RBMs with both the units like these once for doing the means, and then these units with the three range interaction for modeling covariances, you got 69%; as long as you use a lot of these factors. And if you then learn an extra hidden layer of 8,000 units--so now it's just that times that is a hundred million, so there's an extra hundred million connections you learn there. But that's fine because it's unsupervised then you just learn it on lots of data. You get up to 72%. And that's the best result so far on this database. One final thing, you can take this model that was develop for image patches and the student that'll be doing framing recognition just took that code and applied it to log spectrograms, which is sort of more close to what they would like to see, you're not using all these mark up fool stuff, which is designed to throw away stuff you think you don't need and get rid of lots of correlations. Instead you're going to take data that has lots of correlations in but we got a model that can deal with that stuff now. And the first thing George tried on February the 20th, which was four layers of a thousand hidden units on top of this, he got 22.7 correct--percent correct; which was the record for phoneme recognition on the TIMIT database where your not trying to do a model adapted to each speaker. And then a week later when he did that to TIMIT and use more frames, he was down to 21.6%. So this--all this stuff was designed to do vision. It wasn't designed to do phonemes. And if we treat phoneme recognition, it's just a vision problem on a lot of spectral ground. We can wipe out the speech class, at least on small vocabulary. Another student is now, at Microsoft, is seeing if this will work on big recovery as well. >> [INDISTINCT] >> HINTONS: Yes. Yes, right. >> We can give them new better tools. >> HINTONS: We can give them new and better tools. So here's phoneme recognition over the years. Backprop from the 80's got 26.1 percent correct. Over the next 20 years or so, they got that down to 24.4 percent, using methods that weren't learning-inspired so we'll call them artificial. We then got down to 21.6 percent; an estimate of human performance is about 15 percent. I don't know much about how they did this estimate, I'm afraid. But we're about--we're nearly a third of the way from artificial TIMIT. And so we need two more ideas and we're there. Okay, I'm done. I'm finished. >> Questions? >> HINTONS: Yes? >> You mentioned YouTube recently announced that the [INDISTINCT] have broken the world record on the end list of data sets of phoneme recognition by simply using a seven layered feed forward network trained with backprop, but doing it on a GPU with lots and lots of cycles. >> HINTONS: Yes, he did indeed announce that. What he didn't announce was--he's got a spectacular result. He gets timed to 35 errors. What he didn't announce was there's two tricks involved. One trick is to use a big net with lot of layers in a GPU board. That trick by itself wouldn't give you 35 errors. There's a second trick which was sort of pioneered by people at Microsoft in fact, which is to put a lots of work into producing distortions of the data so you have lots and lots of labeled data. So you take a labeled image of a two and you distort it in clever ways and make it still look like a two but be translated so people can then get down to about 40 errors. >> I think they patented that already. >> HINTONS: Good. So Dick's already patented that. So you get down to--you can get down to by 40 errors by doing these distortions. What he did was even better distortions, or more of them, and a much bigger net on a GPU and he got from a 40 to 35, which is impressive because it is hard to make any progress there. But it won't work unless you have a lot of labeled data. And what's--the disguised thing is the work went into--if you look in the paper, it's always straightforward, its just backprop, except when you get to the section of how they generated all those sector labeled data where there's very careful things, like if it's a one or a seven they'd only rotate it a certain number of degrees but if it's something else they rotate it in more degrees. I'm actually the referee for this paper but I don't mind him knowing. I think it's a very important work. But he should emphasize that they have to have labeled data to do that, and they have to put work into distortions. So for me the lesson of that paper is when we small computers, you should put your effort into things like weight constraints so you don't have too many parameters because you only got a small computer. As computer gets bigger and faster, you can transfer your effort from, instead of tying the weights together, like Yann was doing in the early days, put your effort into generating more distortions so you can inject your prior knowledge in the form of distortions and that's much less complication-efficient over the big computers, it's fine and it's more flexible. So I think that's the lesson of that paper. >> I shouldn't even need to ask you a question, you answered it. Thank you. >> HINTON: Any other long question? >> It seems like you've invented some kind of a cortex here that house you expect the property that if it does vision it'll do sound. >> HINTONS: Yes. >> What other problems you going to apply it to? >> HINTONS: Maybe it'd be quicker to say the problems we're not going apply. >> Okay. >> HINTONS: I can't think of any. I mean--okay, let me say what the main limitation of this is for vision. We got at least 10 billion neurons for doing visual things; or at least a billion anyway, probably, 10 billion. And even if we got that many neurons and about 10 to the 13 connections for doing vision, we still have a retina that's got a very small phoneme the size of my thumb there at arms length. And so we still take almost everything and don't look at it. I mean, the essence of vision is not to look at almost everything intelligently; and that's why you got all this funny illusions where you don't see things. We have to do that in these models. These models are completely crazy. And all of computer visions are completely crazy, almost all of it. Because they take a uniform resolution image, and quite a big one like a thousand by thousand, and they try and deal with it all at once with filters all over the image. And if they going to do a selection, they either do it by running off their face to get to everywhere, with no intelligence, or they do sort of interest point detection at a very low level to decide what to attend to. What we do is we fixate somewhere. Then on the basis of what our retina gives us, with these big pixels around the edges and small pixels in the middle, we sort of decide what we seeing and where to look next and by the second or third fixation we've fixating very intelligently and the essence of it is that vision is sampling, it's not processing everything; and that's completely missing from what I said. Now in order to do that, you have to be able to do take what you saw and where you saw it and combined them and that will multiply. So this module, it can do multiply. It's very good in combining what's and where's, to integrate information at a time. And that's one of the things, we're working on that. But that's probably the biggest thing missing. But that is an example of having a module is quite good but now it's never good enough, so you have to put it together over time and use it many times. And that's what sequential reasoning in all this stuff are. So basically, as soon people become sequential we're not modeling that at all. We're modeling what you can do in hundred milliseconds. And so that's what's missing. But I believe that to model that sequential stuff we need to understand what is the sequence of, is the sequence of these very powerful operations. And we're in a better shape now to try and model sequential AI, than we were if we didn't know what a primitive operation is. So this sort of primitive operation was just deciding whether two symbols are the same. We're going to be out of luck for understanding how people do sequential stuff. Yeah. >> This is a [INDISTINCT] question as he said he wanted to do everything if it connects. Are you going to do [INDISTINCT] logic like there exists a God and every girl has a boy she loves? >> HINTON: Hang on, I'm still processing that. Right. Right, I'm making the point that people find "quantifies" quite difficult. >> Oh, yeah. If you [INDISTINCT] quantifiers... >> HINTON: I would love to do that. I have not got a clue how to do it. And you will notice that in old-fashioned AI that you used to point out to [INDISTINCT] people, then you can't do quantifiers, so forget it. Nowadays, when they all do graphical models, they didn't mention that anymore because the graphical models have difficulty of it too. Some people has got [INDISTINCT] some people do. Right. Yeah, some people do. But most of the graphical models of, like, five years ago, they do quantifiers either. And so, a pretty good division line would be what you can do without having to deal with really sophisticated problems like that. I would love to know how we deal with that, but I don't. >> Thank you. >> HINTON: So, yeah, I'm going to give up on that right now. >> So if you had 80 million labeled images and no extra unlabeled ones, would you do your pre-training... >> HINTON: Yes. Yes >> ...and then fine tuning to make us better? >> HINTON: In TIMIT, that's what we have. In TIMIT, all the examples we have labels. It stirs a big wind to do the pre-training. >> But you didn't sneak this result I'm just hearing about? It seems to suggest... >> HINTON: Well, the audience switched it but I haven't tried with all these distortions during pre-training. Now, I've assumed student called [INDISTINCT] who just produced a thesis. Well, he tries things like that. He tries distortions in earnest and he uses special distortions of his own. And the fact is distortions helped a lot. But if you do pre-training, that helps some more too. And [INDISTINCT] results, yes, [INDISTINCT] results, suggest that pre-training will get you through different part of the space even if you have all these labeled data. So clearly, one thing that needs to be done is to try the pre-training and combine with these labels. You don't have to have the pre-training, but I bet you, it still helps. And I bet you, it's more efficient too. It's faster because the pre-training is rather pretty fast, you always have to learn a very good model. You got lots of its features. And starting from there, I think, you'll do better than he does just started from random, and faster. That's just a prediction. You might even get done to 34 out of this. The problem with [INDISTINCT] you can't get significance. TIMIT is really nice that way. They designed it well, so you get higher rates. So you can see differences. >> On the time series aspect, did you see anything that would you get inferences or alterations that are beyond the size of the time window you're using? >> HINTON: Sorry, I didn't understand the question. We have a limited time. We don't... >> You have limited time, after training is there anything of a model that picks up... >> HINTON: Nothing. >> Nothing. >> HINTON: Nothing. It cannot deal with--it can't model host... >> It has an internal state. It has an internal state. >> HINTON: Right. But if sort of what happened 15 times steps ago really tells you what should happen now, and it only tells you what you should happen now. It doesn't tell you what should happen in TIMIT 14 times steps. It just contains information across 15 times steps without having a signature of smaller time scales. You can't pick up on that. >> Okay. >> HINTON: Because it's not got a hidden forward-backward algorithm. A forward-backward algorithm potentially could pick up a lot of load, actually can't. >> So this one wouldn't pick up on things like object permanence or all rules behind the box and comes out of the other side and they're not going to be able to... >> HINTON: Not over a long time scale, no, no. Unless you say that there's a memory involved when you go back to a previous--it gets more complicated, right? Now, it is true that when you build the multilevel one, which you can do with the three interconnections as well as with the three-way connections, at every level you're getting a bigger time span because your time window, it's going further back into the past with each level. So you get a bit high, but that's just sort of linear. >> Can you say--do you have any rules of thumb of how much unlabeled data you need to train each of the different levels and how it would change, like, is it just linear with the number of rates or as you go up levels the things changed? >> HINTON: I have one sort of important thing to say about that, which is that if you're modeling high-dimensional data and you're trying to build an unsupervised model of the data, you need many less trainings on [INDISTINCT] than you would have thought if you use the discriminative learning. When you're doing discriminative learning, there's typically a very few bits per training case to constrain the parameters. You're going to constrain--you got many new parameters for a training case is the number of bits it takes to specify the answer, not the number it takes to specify the input. So within this, you get 3.3 bits per case. If you're modeling the image, the number of bits per case is the number of bits it takes to specify to image which is about a hundred bits. So you need far fewer cases per parameter. In other words what I'm saying is you're modeling much which are things, and so each case is giving you much more information. So actually, we can typically model many more parameters than we have training cases. And discriminative people aren't used to that. Many less parameters than we have pixels and many more than training cases. And in fact, he used about two million cases for doing the image stuff, and it wasn't enough, it was over fitting. He should have used more. But he was fitting 100 million parameters. But the--basically, the only rule of thumb is many less parameters and the number of total number of pixels in your training data, but you can typically use many more parameters in the number of training cases. And you can't do that with normal discriminative learning. Now, if you do do that, when you start discriminative training, it quickly improves things and then very quickly over fits. So you have to stop it early. Okay. >> Okay? >> HINTON: Thanks. >> Let's thank the speaker again. >> Thank you.
B1 中級 米 ディープラーニングの最近の動向 (Recent Developments in Deep Learning) 174 14 Howard Liu に公開 2021 年 01 月 14 日 シェア シェア 保存 報告 動画の中の単語