Placeholder Image

字幕表 動画を再生する

  • - Hello, welcome to a video tutorial.

  • That's what happens on this channel, I guess.

  • So this is sponsored by Spell.

  • Thank you so much to Spell for the sponsorship.

  • What you're about to watch is an edited version

  • of a livestream that happened a couple weeks ago.

  • We have a guest educator and artist,

  • Brooklyn based educator and artist Nabil Hassein.

  • I recommend you check out his website

  • linked in this video's description and learn more

  • about his background and his current work,

  • and all sorts of wonderful stuff that he is up to.

  • So what you're going to see, from beginning to end

  • in this video, is the process

  • for taking a corpus of text,

  • training a machine learning model,

  • this particular model is called LSTM,

  • long short term memory neural network.

  • Nabil will explain that a bit more in the video

  • and offer you some resources to learn about it.

  • Train a model to learn about that text.

  • Train it in the cloud, on Spell,

  • you go to spell.run slash coding train

  • if you want to sign up for that service

  • and follow along with the tutorial.

  • And then download the train model,

  • then bring that train model into the browser,

  • into JavaScript, generate new text in the style of

  • the original text that the model was trained on.

  • So you're going to see the full process for this tutorial.

  • Probably, if you've never watched

  • any of my videos before you're new to coding,

  • you might want to watch some of my workflow videos

  • that show you how to set up your environment

  • you're going to need, you'll need a Python environment,

  • you're going to need a code editor

  • and know how to run a webpage in your browser

  • that you're developing locally on your computer.

  • But I have videos that show all that stuff.

  • I also have a video that introduces the Spell platform

  • and gives you some background about how that works.

  • Alright, so I hope you enjoy this video.

  • If you make something with this, please share it with me.

  • I would love to see what kind of crazy, and interesting,

  • and wacky, and original, and fun, and playful projects

  • you are inspired to make by learning how to do this.

  • Thank you again to Nabil for being here to make

  • this tutorial and to Spell for the sponsorship.

  • Okay, bye bye.

  • - Alright, hello everyone, I'm Nabil,

  • thanks Dan for this great intro

  • and thank Spell for paying me to make

  • this video or to do this livestream.

  • So I have here kind of an outline

  • of what I plan to go through,

  • so I guess I'll start by going ahead

  • and introducing myself.

  • So I already said hi, I'm Nabil,

  • I live in Brooklyn, I'm a freelance

  • technologist, educator, do some other things.

  • Again, thank you Spell for sponsoring this video.

  • So this livestream is about how to train

  • an LSTM model using the Spell platform,

  • so on some remote machine somewhere,

  • and then how to use that model that we've trained

  • using a library called ml5.js,

  • which is a browser based front end library

  • for using machine learning models.

  • So what I'm going to do in this video,

  • I've practiced most of this,

  • I'm going to try to do a few things

  • truly live for you here today.

  • I'm going to kind of extend a project that I did

  • actually at the School for Poetic Computation,

  • which Dan mentioned last summer.

  • The way that that project works is there's a bunch

  • of random, it'll generate random rhymes.

  • Right now, this, what I have live on the web,

  • what I'm actually showing from my website

  • is based on a Markov model,

  • so it's not really machine learning,

  • it's just probabilistic predicting

  • the next character based on the previous ones.

  • Then you can click this all day

  • and it'll keep coming up with more and more rhymes.

  • The video in general, as you know,

  • is about training an LSTM model using Spell

  • and then using it in the browser

  • via a library called ml5.js.

  • So let's go ahead and get into it.

  • So the next thing, so I'm not really going to talk

  • to you much in this video about the theory

  • of neural networks or what is an LSTM really,

  • but I figure I should probably say something.

  • First of all, LSTM stands for long short term memory.

  • It's a specific type of recurrent neural network,

  • and what is useful about recurrent neural networks

  • or RNNs compared to some other types

  • of neural networks is the way

  • that their architecture includes loops,

  • and that can be useful for kind of

  • keeping data around in the network,

  • so to speak, which is very useful

  • for applications involving

  • natural language, human language,

  • because context matters so much in language.

  • Predicting the next character or the next word,

  • you might get a much better prediction

  • if you actually remember what was said

  • even some while ago, maybe like

  • much earlier in a long sentence.

  • I have a few quick references here,

  • which, by now are a little old,

  • but these are what I read to learn

  • a little bit about recurrent neural networks.

  • So there's this blog post called The Unreasonable

  • Effective of Recurrent Neural Networks,

  • and there's this other blog post

  • called Understanding LSTMs.

  • So yeah, this gives a little bit of overview

  • of kind of the same stuff I was just talking about.

  • Humans don't start their thinking from scratch every second.

  • You understand each word based on

  • your understanding of previous words,

  • and that's what we want our network to do as well,

  • which is why we're going to use this LSTM model.

  • I know that before I had the chance,

  • while preparing for this video,

  • to watch a video that Dan made kind of

  • giving an overview of the Spell platform,

  • so a link that video will also

  • be added to the video description

  • and you can kind of get into a little

  • bit more depth about using Spell.

  • And I'll also mention some things

  • about using Spell as we go through this.

  • Okay, so when you want to do a project like this,

  • the first thing that you have to do

  • is get your corpus of data.

  • So in this case, since I was getting song lyrics,

  • I used a site called Genius.com,

  • which you might be familiar with.

  • It's a popular lyrics website,

  • it has some other features too

  • but the main thing I use it for,

  • and I think most people use it for,

  • is reading lyrics.

  • So what I'm going to do, I'm going to try to do

  • everything kind of from scratch, so to speak,

  • so that you should be able to follow along in theory.

  • What I'm going to do, this is a folder that I used to prepare.

  • What I'm going to do is just make a new folder

  • called Spell Livestream, and I'm going to

  • do everything from inside of this folder,

  • which just lives somewhere on my computer.

  • So right now this folder is empty.

  • And so the first thing that I'm going to do is just

  • clone my generative DOOM repository from GitHub.

  • There's only actually one file in there

  • that I care about so maybe not actually clone

  • the whole repository, let me just get that one file.

  • Okay, so I'm just going to

  • where this is, it's in data, oh but did I push it up?

  • I have so many branches here.

  • Okay, why don't I use the one that I have on my computer.

  • So I'm just going to copy a file

  • that I have on my computer into this folder.

  • So where's that, in Spell demo slash

  • generative DOOM slash data.

  • I have a file called input.txt that I just moved,

  • that I just brought a copy of into my current directory.

  • We can just check it out really quickly,

  • oops, less input.txt.

  • So you can see this is just the list of lyrics.

  • Okay, this is my corpus.

  • It's worth noting that the data set

  • I'm actually using for this example isn't that big.

  • We can check the size of it with the command

  • line utility du for this usage,

  • past the human readable flag so that we can actually

  • tell how big this file is.

  • It's about 308 kilobytes, so it's not huge.

  • Normally when you're training machine learning models,

  • kind of, the more data the better,

  • but I got all the lyrics I could find by this artist,

  • this is really the most that I could get.

  • So that's what we're going to use.

  • Cool, so it's also worth noting

  • that you have to clean the data before you train it,

  • so I can actually go ahead and show

  • the code that I used to get these lyrics.

  • I'm not going to go into full depth, but again,

  • it's on my GitHub if you want to check it out.

  • So let's put it over here.

  • So I happen to do my scripting using Python.

  • You can do this in any language,

  • you can do web scripting using Node.js

  • or Ruby or whatever your favorite language is.

  • But I happen to have already used before

  • a Python library called BeautifulSoup,

  • which is very useful for web scripting.

  • It so happens that Genius.com happens to keep

  • their lyrics and their URLs follow a pattern like this,

  • genius.com slash the name of the artist,

  • which I substituted in here,

  • and the name of the song,

  • which I substituted in here,

  • and then I used another Python library

  • called Requests to just go ahead

  • and fetch all these different things.

  • So this is the basic idea,

  • I'm not going to go into full depth,

  • but I just kind of hard coded a lot of names

  • of different songs into here,

  • and then I have a main loop which basically

  • just loops through each artist's name

  • because DOOM has actually recorded

  • under many different names, so I can't just

  • use the same artist's name all the time.

  • And then the same thing for the albums

  • and then finally the songs in order

  • to go ahead and just fetch all of this data.

  • The thing is that when you just go directly to some

  • lyrics website, like when you fetch

  • the data on the page, you end up getting

  • a lot of other data that you don't

  • really care about in the HTML,

  • and so an important step is to clean the data

  • so that when you're training the model

  • you're only training it on the actual

  • corpus that you care about

  • and you're not training it on the angle brackets

  • of HTML tags or something like that

  • that you don't actually want.

  • So again, I think I have most of the code

  • that I used to clean it on the GitHub,

  • I think it is there, but if not there are

  • other resources that you can use

  • to learn more about data cleaning.

  • Again this video is really about

  • training machine learning models using Spell

  • and then using them in the browser.

  • So let's get back to that.

  • I wanted to mention Project Gutenberg

  • is another resource that has lots of free text

  • that's in the public domain that you can just use.

  • Web scraping with Node.js is another resource

  • that I've happened to look at

  • for doing this kind of thing.

  • And so although my script.py file in my generative

  • DOOM repository doesn't show this,

  • the original version I kind of kept each file,

  • kept each song, in it's own file of lyrics.

  • But it so happens that the machine,

  • where I'm going to show you next,

  • works with an input that's just one big input file,

  • input.txt, so I just did some boring shell script

  • stuff that just concatenate all the files together.

  • And I've already noted that my dataset is kind of small.

  • That's everything that I wanted to say about getting there.

  • So let's kind of get into the main thing.

  • So I think I already did this part,

  • I created a new directory for all

  • this stuff to live in.

  • Okay.

  • So let's go ahead and go on to the next step.

  • So it's a good habit, I think, to use virtualenv.

  • I use it with Python two, I understand

  • things have kind of moved on a bit with Python three,

  • but I'm still on the Python two,

  • so I'm going to use this virtualenv thing

  • to kind of keep my dependencies isolated.

  • Although I think there should actually only be one.

  • But let's go ahead and do that anyway.

  • So I have some other virtualenv active right now,

  • it so happens, I see that from my command prompts over here.

  • So I just ran this command deactivate to get rid of that.

  • I'm just going to clear the screen

  • to make it a little less noisy here.

  • And then what did I want to do?

  • I wanted to create a new virtualenv.

  • And what did I want to call it?

  • I want to call it spell video virtualenv.

  • Okay.

  • So it's setting up a new virtual environment,

  • which lives inside of this folder here.

  • And the way that you use Python virtualenv

  • to get it active is you say, what is it,

  • spell video virtualenv slash bin slash activate,

  • we'll have to say source at the beginning.

  • Source and then the path to this activate script.

  • Okay, alright.

  • And now you can see my prompt changed

  • because I happen to have my terminal preferences

  • set up that way so that I can

  • remember what virtualenv I'm in.

  • Okay, so I did that.

  • Oh yeah, I already went and got that input file,

  • which I should probably push it up,

  • I haven't actually pushed up to the GitHub,

  • like the one file version, but you know,

  • life is messy.

  • But what I am going to get the proper version

  • of from GitHub is this repository

  • called training LSTM and so I'm just

  • going to go ahead and clone that,

  • and let's actually go and take a look

  • at that repository and its read me.

  • Cool, so you can see that this,

  • you can see from the description of this

  • repository training in LSTM network

  • and using the model ml5.js that it's

  • highly relevant to what we're doing in this video.

  • The directions in this repository's read me

  • are based on what you would want to do

  • if you were training

  • the machine learning model locally on your own computer.

  • So if that's what you want to do,

  • you can go ahead and follow this,

  • but since this video is about how to use Spell,

  • that's what I'm actually going to do,

  • so I'm not going to follow these directions exactly,

  • but we are going to follow along

  • with certain parts of it.

  • Okay, so I've already cloned this repository,

  • right, that was the last command that I ran,

  • so I'm going to go ahead and enter into that repository.

  • And then what I want to do is create

  • a directory inside of here called data.

  • And then I'm going to go ahead and move that

  • input.txt file into that data directory.

  • Or a copy, I'd rather.

  • I guess I could've deleted it in

  • the other directory, but whatever.

  • Okay.

  • Okay, great, so this is the setup.

  • We have a repository.

  • We have this repository locally that is going to

  • help us train an LSTM network using TensorFlow.

  • And then we're going to, after we train the model,

  • we can use it in ml5.js.

  • So we're pretty much done with our setup,

  • let's get into actually training the model.

  • So again, this is the link that you can use

  • to sign up for Spell if you haven't already.

  • It so happens that I have already,

  • so I should be logged in here,

  • let me just make sure I haven't been logged out.

  • I haven't.

  • So I'm in here in Spell.run,

  • and it gives me some information

  • about how to install Spell,

  • how to log-in with Spell,

  • there's a little quick step guide

  • that you can check out with some of the resources

  • that I used when preparing for this video.

  • So yeah, like I mentioned, that other

  • training LSTM repositories tells you how to run locally,

  • but for us all we really do need to install is Spell,

  • so I'm going to go ahead and do that with pip install Spell.

  • And it's going to go ahead and fetch Spell

  • and whatever things Spell depends on

  • from the Python packing, Py Py, whatever it's called,

  • it's going to just go ahead and get that.

  • Okay.

  • And then once it's done installing I'll be able to log in.

  • Alright, I can remember it.

  • So, you see it greeted me, hello Nabil Hassein,

  • so that's me, so I am logged in as myself.

  • And if you ever forget who you're

  • logged in as for some reason,

  • the Spell command has a lot of sub commands,

  • like Spell who am I will tell you who you are.

  • I'm just going to go ahead and get started

  • with training this model,

  • and the first thing that we need to do

  • is to upload the file to Spell, okay.

  • So what I want to run is this command here.

  • Spell upload the path on my local computer of the file,

  • and then I want to give it a name

  • of where I'm going to upload it to.

  • Okay, so I just got to place that command.

  • Spell upload my data slash input.txt to this destination.

  • Oh it's actually going to prompt me

  • for the destination in a minute.

  • It doesn't want that yet but it's

  • going to want that momentarily.

  • So let me just say spell upload

  • the name of the file that I want to upload.

  • And now it's asking me for the name of the upload

  • that I was trying to give a little bit early.

  • And it tells me that the file,

  • this is the path to it on my local computer

  • that I'm typing on right now will be accessible

  • at upload slash name slash input.txt.

  • And that's the name, oops, oops,

  • just part of it is the name

  • that I want to put in so I'm just going to

  • delete the part I don't want and put in,

  • what was it, Nabil spell livestream DOOM.

  • Okay.

  • Total upload size, 307k, same as what we saw,

  • or at least very close to what we saw

  • when we ran the du command earlier.

  • And the upload completed.

  • So that's great.

  • So what we're going to do now,

  • this is kind of the most,

  • probably most complicated command that

  • we're going to run but it's really the only one.

  • This is really what's saying to go ahead

  • and actually run the Python script

  • that we downloaded from train,

  • that training LSTM git repository.

  • Okay, we're going to run that with Python,

  • this train.py script with the data dir set too,

  • what I happen to call data,

  • that the name of the folder where I put that input.txt.

  • And I'm going to run it and I'm going to mount the data,

  • I'm going to mount the folder that I had

  • just created by uploading that file as the name data

  • so that it can understand this data directory.

  • Okay, so I should get, I think,

  • one error when I do this, and I'll talk about why.

  • Okay, spell run mount dash dash mount upload

  • slash Nabil Spell DOOM, ope, I called it livestream,

  • in my notes I didn't quite update

  • from when I practiced so let me go ahead and fix that.

  • Okay, let's see.

  • Okay, so let's try that.

  • And it tells me that there's some

  • untracked files in this repository

  • and they won't be available on this run.

  • Continue the run anyway?

  • No, that's really the file that I care about.

  • So Spell encourages us to use the Git

  • version control system to make sure

  • that the data that we're training on is checked in,

  • and that's very good for reproducibility,

  • if we want to go back later and understand

  • what was going on and what was there,

  • so I'm going to go ahead and kind of follow this

  • adjusted workflow so I'm going to go ahead

  • and git add dot, or I guess I could

  • git add data, whatever.

  • That's the only thing, same effect

  • in this particular case, and I'm going to

  • say add data dot input.txt file of DOOM lyrics.

  • Okay.

  • And now, having done that,

  • if I run the same command it won't give me

  • that same warning since the files

  • are now tracked instead of being untracked.

  • Okay.

  • Let me go ahead and start to do this

  • and then I'll go ahead and mention that other thing

  • that I said I was going to mention.

  • So I'm just going to press up a few times

  • and just go back to my history

  • to run the same command again.

  • Spell run, I'm mounting that data folder

  • that I uploaded to be called data.

  • Oh I put it in the wrong order.

  • - [Dan] So one thing to mention, I think,

  • is that if you're going to mount the data file,

  • you don't actually have to commit it.

  • - Oh.

  • 'Cause otherwise if you pull it

  • from there you can do either.

  • - Oh, I see.

  • - [Dan] So if you don't want to upload it with git,

  • then you can do the mounting thing

  • that you're showing now.

  • - Okay, yeah, oh right, so I'm kind of,

  • so there's more than one way to work with Spell,

  • and I think I kind of conflated

  • two of them a little bit.

  • So, yeah, I didn't actually need to do

  • the git commit because of the way

  • that I'm doing this because I uploaded it before,

  • and that will also kind of give some of the same,

  • reproducibility benefits because Spell will

  • keep a record of what we uploaded,

  • but it doesn't hurt to get committed either.

  • Let me just fix that typo.

  • - [Dan] In this case, it's such

  • a small file it doesn't really matter.

  • - Yeah.

  • - [Dan] If you were working with a huge

  • gigabyte file or something,

  • you'd want to upload that separately

  • without having to commit it also.

  • - Right, yeah, because Git isn't always

  • the best for dealing with large files

  • which is why there's tools like Spell

  • and tools like, what is it, Git,

  • whatever, there's a bunch of tools.

  • Cool, so let me just go ahead and fix that typo.

  • I should fix that on my notes as well.

  • Tells me everything is up to date

  • because I did make, like, a commit,

  • although, like Dan mentioned,

  • I didn't really have to.

  • Tells me it's casting spell number 16,

  • so I happen to have used Spell about 15 times before.

  • Tells me I can stop viewing the logs

  • with control C, tells me it's requesting

  • a machine and building and mounting,

  • and the run is running.

  • And, so, this is still running,

  • and like it told me before,

  • I can just get out of the logs in my

  • local terminal with control C,

  • but this is running on a remote machine

  • that Spell has provisioned and set up very nicely

  • for me so I don't have to worry about it.

  • So I'm not actually stopping the run

  • from happening when I control C,

  • I'm just stopping the logs

  • from appearing in my own terminal.

  • If I want to check out those logs,

  • I can say Spell logs 16,

  • and they'll start appearing again.

  • And there's also some other commands

  • that it told me about, like I could kill it

  • with Spell kill whatever, but I don't want to,

  • I'm going to let it keep running.

  • And besides checking out the logs locally

  • with Spell logs the number of the run,

  • you could also come over here to this Spell Web UI,

  • and check out different information

  • about the run in here.

  • But as you may notice from this video,

  • I tend to have a preference for the command lines,

  • so I'm going to keep doing things mostly that way.

  • Cool.

  • So let's see.

  • Oh yeah, so one thing I did want to

  • mention was the parameters,

  • or what is sometimes called the hyperparameters

  • of the network, so let's just go back

  • to this git read me really quick.

  • So yeah, like I said, this gives you more

  • information about how you would run it locally,

  • including how you can pass additional flags

  • that I didn't bother passing to control

  • more of the characteristics of the network,

  • like its size, how many layers they are built

  • into the sequence, and various other things

  • that you can read more about in this repository.

  • They have here some recommendations for what

  • you might want to select for your hyperparameters

  • according to the size of your training dataset.

  • Because my training dataset is so small,

  • I decided the defaults were probably fine.

  • The next thing that I wanted to talk about

  • was the difference between running

  • on CPU versus GPU.

  • So I imagine this might be review for many viewers,

  • but I am a teacher, so I'm always

  • a fan of just reviewing material.

  • So the CPU is the central processing unit

  • of your computer that has maybe a little

  • bit of parallelism, but for the most part

  • is best at running things sequentially, very fast.

  • And the model of computation is a little

  • bit different from that of GPU,

  • which at some point stood for graphics

  • processing unit and maybe still does,

  • but maybe that acronym's been retired by now

  • because GPUs actually have very many applications

  • other than graphics, including training

  • neural networks on text, for example.

  • GPUs, historically they got, I think,

  • that name because each pixel on a screen

  • for the most part is independent of each other one

  • and so can be computed independently,

  • and so a GPU is much more highly

  • parallel compared to a CPU.

  • It's not going to be as fast at

  • completing like one single task,

  • but it is very good for displaying things

  • on screens and it also happens to be

  • very good for training neural networks.

  • So in the last command that I ran over here

  • on the command line to train the neural network,

  • this is running by a CPU,

  • and what I could do if I wanted to

  • instead run my code on a GPU

  • is just tell it that that's the type of machine

  • I want by adding this dash dash machine type flag.

  • And the machine type that I'm going to use is K80,

  • okay, so where does this K80 come from?

  • Well, if you check out spell.run slash pricing,

  • you'll see some information about how much

  • Spell charges for different types of machine types

  • according to your needs for whether this GPU

  • or that CPU is the best for your particular task.

  • As you can see there's a few different ones,

  • CPU, CPU big, you know, all these things, K80.

  • K80 happens to be one of the less expensive GPU units,

  • so it's good enough for my purses

  • and that is what I'm going to use.

  • Okay, so I go ahead and run that command.

  • Everything's up to date, it's casting spell number 17,

  • we see a very similar bit of output as we did before

  • as it gets ready to start training the model.

  • Okay, the run is running, so in a moment we should

  • start seeing the logs, after it reads the text file.

  • Alright.

  • So now this run is running,

  • and I don't know how obvious this is

  • if you're following along, but it's actually

  • noticeable at least for me that

  • this is happening a lot faster.

  • This model is being trained a lot

  • faster via the GPU than the CPU one was.

  • So the CPU one got a head start,

  • but I still expect that the GPU one

  • will actually finish substantially faster.

  • We see we're already at like 413 out of 6,000

  • iterations that it's going through to train the model.

  • Like if we check in on the previous one,

  • let's see how far it is.

  • Yeah, actually, okay, no, I mean

  • the head start it had was pretty big,

  • but you can see the GPU one is moving faster.

  • Like, if we actually go in,

  • because like I mentioned before,

  • I've had a few practice runs here before,

  • we can look at a few ones that I did before.

  • Yeah, these were kind of my first two practice runs.

  • Using a subtly different model, you can see here,

  • but this one on CPU took close to five hours,

  • and on GPU it took only a little

  • bit more than 15 minutes.

  • Yeah, so GPU is faster for this particular use case.

  • Okay, so just for the sake of time,

  • I'm going to grab a model that I had already

  • trained before rather than just waiting for this one

  • to go all the way through, although we could.

  • We could do that, I mean we can actually use

  • this model later if people are interested in that.

  • But so what I want to do to grab the data

  • from Spell is to run this command here,

  • spell cp runs slash the number of the run slash models.

  • Okay, that's how I'm going to fetch that data.

  • Okay, so I'm just going to cd up here,

  • now I'm kind of in my home folder of all

  • the different things that I'm kind of

  • grabbing from here, grabbing from there

  • to put together into this demo.

  • I'm going to go ahead and run spell cp, it was 15,

  • right, let me just look here again.

  • Yeah, yeah, so you can see this is using

  • that same training LSTM I was talking about,

  • it completed in about five minutes.

  • Actually I guess this one should

  • complete pretty quickly, too.

  • And that was a practice run that I did just

  • a few minutes before this livestream started.

  • So I'm going to spell cp runs slash 15 slash models.

  • And it's copying 10 files that are from there.

  • My ls, my ls data, it remembers the same data

  • directory that I passed in before as the name.

  • And these 10 files constitute the model, okay,

  • I'm not really going to go into depth about

  • what are these files and what's in them,

  • but yeah, if you're following along

  • you could poke into them and check it out.

  • Cool, so we've trained the model.

  • We've used Spell to train a LSTM model

  • on a corpus of data that we obtained,

  • and now that we have the model,

  • let's use it for something.

  • So I'm going to borrow and then modify

  • an example from this repository here,

  • on the ml5.js GitHub account they have

  • a repository called ml5 examples.

  • So there's a whole collection of examples,

  • there's a bunch of them.

  • You can find out a little bit about how

  • it's organized and some other stuff from their read me.

  • I'm going to use one in the p5js folder.

  • We are worried about LSTMs.

  • Yeah, that interactive one is also interesting.

  • That's more

  • I mean, it's more interactive,

  • I'm not really going to describe it.

  • We're going to use the non-interactive version, LSTM texts.

  • And we have here just a few files.

  • So they actually have a pre-trained model

  • that I'm going to just ignore and not use

  • because we're going to use our model

  • that we just trained instead,

  • but what I am going to do is just fetch

  • these two files, this HTML file and sketch.js file.

  • And because this repository's big

  • and I just don't want to wait to clone it,

  • I'm literally just going to fetch

  • these two files and nothing else.

  • Okay.

  • So what I'm going to do is just,

  • we'll create another directory, why not.

  • Ml5 LSTM example,

  • and I will change my current directory to be in there.

  • Let me just clear my screen for clarity's sake,

  • and then I'm just going to use the command

  • line program wget, which will just fetch the raw file,

  • I did have to click raw on GitHub.

  • And we'll fetch it onto my local machine.

  • So I do that,

  • and then I go back and I do the same thing with sketch.js.

  • I just find this one raw file,

  • copy the URL and I use the program wget

  • to download it locally.

  • Okay, so now if I list what's here,

  • I have these two files, index.html and sketch.js.

  • So let's take a minute to check out,

  • we'll read the files themselves

  • and we'll also use them.

  • So what I'm going to do is just run

  • a program called http server,

  • which you could install if you want,

  • if you don't already have it with,

  • what is it, mpm install dash g http dash server.

  • You can use any of it, if you're used

  • to using a different web server,

  • anything that will serve up index.html

  • in your local folder is fine.

  • So it tells me where, tells me

  • the URL I can go to on local host

  • to check this out so I'm going to go there.

  • Says LSTM Text Generation Example,

  • this example uses a pre-trained model

  • on a corpus of Virginia Wolf,

  • although I'm actually not doing that,

  • so I might change that.

  • So let's actually go ahead and go into this file

  • and also look at the JavaScript file.

  • So let me, I'm a iMAX user,

  • I'm just going to go ahead and open up repos

  • and I call it Spell livestream.

  • So these are the two files that I had

  • just downloaded a moment ago.

  • It's index.html and sketch.js,

  • so let me open those up in a way

  • that's going to be a little bit more readable for you.

  • My notes before were so, I guess

  • let me do it over here then.

  • The folders don't really bother

  • the video as much as this one does.

  • Okay, so we have here an HTML document,

  • which relies on p5

  • and ml5 as the libraries that are being used

  • and pretty much nothing else.

  • Alright, so this example uses

  • a pre-trained model on a corpus of MF DOOM.

  • So let's just make this nice and accurate.

  • Okay.

  • It says the meaning of life is,

  • which isn't something I remember DOOM saying,

  • but whatever, we can leave that

  • as the seed text for now.

  • That's an input field in HTML,

  • so we can just change that anyway.

  • So we have a few sliders for how long

  • we want the output to be and the temperature,

  • which we'll talk about more a little later.

  • And what's really interesting is the sketch.js file,

  • so let's actually take a look there.

  • It says open source software,

  • I can just do whatever I want with it,

  • which is great.

  • Okay, so we declare a few variables here,

  • so again, this video isn't about p5,

  • there's kind of a lot of things

  • that I'm touching on but not getting into,

  • but p5 is a really cool library,

  • I encourage you to check that out

  • if you're not familiar with it already.

  • It's great for a lot of artistic coding,

  • I've used it for some other projects as well.

  • There's kind of two main functions in,

  • well, yeah, let me not really get into p5.

  • We're going to start with the set up function.

  • Now I'll just say that p5 will run

  • this set up function for us at the beginning.

  • And it says create the LSTM generator

  • passing it the model directory.

  • I don't have anything called model slash wolf

  • 'cause I didn't clone that whole repository,

  • so what I need to do is make sure that this,

  • when it's generating, like when it's creating

  • the LSTM that we're going to use,

  • I need to make sure that this is pointing

  • towards the proper path where we have our file.

  • So let me go ahead and just remind myself on the command

  • line of where I am keeping everything.

  • Let me just control C out of the server

  • and I'll start again in a minute.

  • So let's see.

  • Let me ls up here, I have something called

  • data up there, which had what I wanted, right?

  • Yeah, those are the files from my model.

  • So why don't I just copy that folder into here, where I am.

  • I have to say, what is it, lower case R

  • or capital R for our cursive copy.

  • I guess lower case worked.

  • So now when I ls here, besides the two files

  • that I fetch from GitHub using wget,

  • I also have all those data files that are right here.

  • So what I'm going to do is I'm just

  • going to change this to say data,

  • because that's where my data is, okay.

  • And then here, there's some other code.

  • This is really about the user interaction

  • of like what happens when sliders

  • get moved around and buttons get clicked,

  • so I'm not really going to go over that.

  • What we will just take a minute to look at

  • is this generate function, which again,

  • not going to go all the way through,

  • but it lets you know it's generating stuff,

  • like just so that the user knows something

  • is happening, grabs the input from the seed text.

  • It uses these parameters, temperature and length,

  • and then the ml5.js library does really all

  • the heavy lifting for us,

  • we just kind of cull this lstm.generate function

  • with our data and patch up this call back

  • function which will go ahead and update the dom,

  • kind of update the HTML page when that function

  • is done running, done generating the text

  • that the model has predicted based on the input seed.

  • Okay.

  • So you can see this is a pretty short file.

  • Didn't go through every detail,

  • but it's on GitHub, you can check it out.

  • You saw I really just made one small change to it.

  • Cool, so let me go back to my notes.

  • I'm pretty sure I know what I want to do next,

  • but it's always good to be sure.

  • That's the path, yup, I already told you that.

  • I did that.

  • Alright, so let me actually go ahead

  • and run the server again.

  • Okay, just refresh the page,

  • you can see it's updated with this.

  • Okay, the model's loaded, we can click generate.

  • And it, I mean, I don't know how many

  • people listen to DOOM, if you don't

  • maybe you can just take my word for it.

  • This sounds a little bit like something he might say.

  • So we can adjust the length, make it longer,

  • or make it shorter,

  • and we can use this temperature thing, like,

  • the temperature is something like intuitively,

  • it's kind of like the randomness of the text.

  • The higher the temperature, the less random it will be,

  • kind of the more derivative it will be

  • of the original text, and if you turn it up

  • very high it starts to become very likely

  • that you get kind of direct quotes

  • from the original corpus.

  • If it's lower, it is maybe a little

  • bit more chaotic, so to speak.

  • And I think that can generate things

  • that are a little bit more out there

  • and more interesting or original.

  • But if you start to do it maybe too low,

  • you might start to get kind of just nonsense.

  • It might not really make very much sense at all.

  • Especially if you get like really low.

  • Oh, it might or it might not.

  • So yeah, I mean, okay, I'll withhold

  • my opinion on these generated lyrics for now.

  • This is an art critique session.

  • Alright, so yeah, I mean, that's really the main thing,

  • so I mean if I was going to go ahead

  • and reproduce my original project,

  • what I would do now is pull in another

  • dependency which is the pronouncing library,

  • and then I would've, what I should've done,

  • maybe I can do this now, is trained

  • the model backwards, like if we actually look at

  • this input.txt you can see

  • that these lyrics are forward

  • and so the model does that also,

  • so that's something that I would

  • want to do is just reverse that input

  • and train the model backwards.

  • And then I can use the pronouncing

  • thing to go backwards and so on.

  • But, yeah, I think instead of doing that

  • I think what might make more sense

  • would be for me to take any questions

  • from the people who are on the livestream

  • because that is pretty much, yeah,

  • that's pretty much that.

  • So we've got through what is LSTM,

  • getting data, setting things up,

  • training the model using Spell,

  • and then using the model in ml5.js,

  • so that is what I had prepared for you today,

  • and I look forward to any questions.

  • - Thank you so much Nabil, for this wonderful tutorial.

  • Again, thank you to Spell for the sponsorship.

  • If you make something with this,

  • please share it with me,

  • I would really love to know about it,

  • and I'll see you in future tutorial.

  • Many more machine learning, ml5,

  • Python-y, TensorFlow-y things to come I hope.

  • Goodbye.

  • (upbeat music)

- Hello, welcome to a video tutorial.

字幕と単語

ワンタップで英和辞典検索 単語をクリックすると、意味が表示されます

A2 初級

Nabil HasseinによるLSTMとスペルによるテキスト生成 (Text Generation with LSTM and Spell with Nabil Hassein)

  • 3 0
    林宜悉 に公開 2021 年 01 月 14 日
動画の中の単語