Placeholder Image

字幕表 動画を再生する

  • JARED DUKE: Thanks everybody for showing up.

  • My name is Jared.

  • I'm an engineer on the TensorFlow Lite team.

  • Today I will be giving a very high level overview

  • with a few deep dives into the TensorFlow Lite

  • stack, what it is, why we have it, what it can do for you.

  • Again, this is a very broad topic.

  • So there will be some follow up here.

  • And if you have any questions, feel free to interrupt me.

  • And you know, this is meant to be enlightening for you.

  • But it will be a bit of a whirlwind.

  • So let's get started.

  • First off, I do want to talk about some

  • of the origins of TensorFlow Light

  • and what motivated its creation, why we have it

  • in the first place and we can't just use TensorFlow on devices.

  • I'll briefly review how you actually use TensorFlow Lite.

  • That means how you use the converter.

  • How you use the runtime.

  • And then talk a little bit about performance considerations.

  • How you can get the best performance on device

  • when you're using TensorFlow Lite.

  • OK.

  • Why do you need TensorFlow Lite in your life?

  • Well, again, here's some kind of boilerplate motivation

  • for why we need on device ML.

  • But these are actually important use cases.

  • You don't always have a connection.

  • You can't just always be running inference in the cloud

  • and streaming that to your device.

  • A lot of devices, particularly in developing countries,

  • have restrictions on bandwidth.

  • They can't just be streaming live video

  • to get their selfie segmentation.

  • They want that done locally on their phone.

  • There's issues with latency if you need

  • real time object detection.

  • Streaming to the cloud, again, is problematic.

  • And then there's issues with power.

  • On a mobile device, often the radio

  • is using the most power on your device.

  • So if you can do things locally, particularly with a hardware

  • backend like a DSP or an MPU, you

  • will extend your battery life.

  • But along with mobile ML execution,

  • there are a number of challenges with memory constraints,

  • with the low powered CPUs that we have on mobile devices.

  • There's also a very kind of fragmented and heterogeneous

  • ecosystem of hardware backends.

  • This isn't like the cloud where often

  • you have a primary provider of your acceleration backend

  • with, say, NVIDIA GPUs or TPUs.

  • There's a large class of different kinds

  • of accelerators.

  • And there's a problem with how can we

  • actually leverage all of these.

  • So again, TensorFlow works great on large well-powered devices

  • in the cloud, locally on beefy workstation machines.

  • But TensorFlow Lite is not focused on these cases.

  • It's focused on the edge.

  • So stepping back a bit, we've had TensorFlow

  • for a number of years.

  • And why couldn't we just trim this down

  • and run it on a mobile device?

  • This is actually what we call the TensorFlow mobile project.

  • And we tried this.

  • And after a lot of effort, and a lot of hours,

  • and blood, sweat, and tears, we were

  • able to create kind of a reduced variant of TensorFlow

  • with a reduced operator set and a trimmed down runtime.

  • But we were hitting a lower bound

  • on where we could go in terms of the size of the binary.

  • And there was also issues in how we

  • could make that runtime a bit more extensible,

  • how we could map it onto all these different kinds

  • of accelerators that you get in a mobile environment.

  • And while there have been a lot of improvements

  • in the TensorFlow ecosystem with respect to modularity,

  • it wasn't quite where we needed it

  • to be to make that a reality.

  • AUDIENCE: How small a memory do you need to get to?

  • JARED DUKE: Memory?

  • AUDIENCE: Yeah.

  • Three [INAUDIBLE] seem too much.

  • JARED DUKE: So this is just the binary size.

  • AUDIENCE: Yeah.

  • Yeah.

  • [INAUDIBLE]

  • JARED DUKE: So in app size.

  • In terms of memory, it's highly model dependent.

  • So if you're using a very large model,

  • then you may be required to use lots of memory.

  • But there are different considerations

  • that we've taken into account with TensorFlow Lite

  • to reduce the memory consumption.

  • AUDIENCE: But your size, how small is it?

  • JARED DUKE: With TensorFlow Lite?

  • AUDIENCE: Yeah.

  • JARED DUKE: So the core interpreter runtime

  • is 100 kilobytes.

  • And then with our full set of operators,

  • it's less than a megabyte.

  • So TFMini was a project that shares

  • some of the same origins with TensorFlow Lite.

  • And this was, effectively, a tool chain

  • where you could take your frozen model.

  • You could convert it.

  • And it did some kind of high level operator fusings.

  • And then it would do code gen. And it would kind of

  • bake your model into your actual binary.

  • And then you could run this on your device and deploy it.

  • And it was well-tuned for mobile devices.

  • But again, there are problems with portability

  • when you're baking the model into an actual binary.

  • You can't always stream this from the cloud

  • and rely on this being a secure path.

  • And it's often discouraged.

  • And this is more of a first party solution

  • for a lot of vision-based use cases and not a general purpose

  • solution.

  • So enter TensorFlow Lite.

  • Lightweight machine learning library

  • from all embedded devices.

  • The goals behind this were making ML easier,

  • making it faster, and making the kind of binary size and memory

  • impact smaller.

  • And I'll dive into each of these a bit more in detail

  • in terms of what it looks like in the TensorFlow Lite stack.

  • But again, the chief considerations

  • were reducing the footprint in memory and binary size,

  • making conversion straightforward,

  • having a set of APIs that were focused primarily on inference.

  • So you've already crafted and authored your models.

  • How can you just run and deploy these on a mobile device?

  • And then taking advantage again of mobile-specific hardware

  • like these ARM CPUs, like these DSP and NPUs

  • that are in development.

  • So let's talk about the actual stack.

  • TensorFlow Lite has a converter where

  • you ingest the graph def, the saved model, the frozen graphs.

  • You convert it to a TensorFlow Lite specific model file

  • format.

  • And I'll dig into the specifics there.

  • There's an interpreter for actually executing inference.

  • There's a set of ops.

  • We call it the TensorFlow Lite dialect

  • of operators, which is slightly different than

  • the core TensorFlow operators.

  • And then there's a way to plug in these different hardware

  • accelerators.

  • Just walking through this briefly, again,

  • the converter spits out a TFLite model.

  • You feed it into your runtime.

  • It's got a set of optimized kernels

  • and then some hardware plugins.

  • So let's talk a little bit more about the converter itself

  • and things that are interesting there.

  • It does things like constant folding.

  • It does operator fusing where you're

  • baking the activations and the biased computation

  • into these high level operators like convolution, which

  • we found to provide a pretty substantial speed up

  • on mobile devices.

  • Quantization was one of the chief considerations

  • with developing this converter, supporting

  • both quantization-aware training and post-training quantization.

  • And it was based on flat buffers.

  • So flat buffers are an analog to protobufs, which are

  • used extensively in TensorFlow.

  • But they were developed with more real time considerations

  • in mind, specifically for video games.

  • And the idea is that you can take a flat buffer.

  • You can map it into memory and then read and interpret that

  • directly.

  • There's no unpacking step.

  • And this has a lot of nice advantages.

  • You can actually map this into a page and it's clean.

  • It's not a dirty page.

  • You're not dirtying up your heap.

  • And this is extremely important in mobile environments

  • where you are constrained on memory.

  • And often the app is going in and out of foreground.

  • And there's low memory pressure.

  • And there's also a smaller binary size

  • impact when you use flat buffers relative to protobufs.

  • So the interpreter, again, was built from the ground

  • up with mobile devices in mind.

  • It has fewer dependencies.

  • We try not to depend on really anything at base.

  • We have very few absolute dependencies.

  • I already talked about the binary size here.

  • It's quite a bit smaller than--

  • the minimum binary size we were able to get with TensorFlow

  • Mobile was about three megabytes for just the runtime.

  • And that's without any operators.

  • It was engineered to start up quickly.

  • That's kind of a combination of being able to map your models

  • directly into memory but then also having a static execution

  • plan where there's--

  • during conversion, we basically map out

  • directly what the sequence of nodes that would be executed.

  • And then for the memory planning,

  • basically there's a pass when you're running your model where

  • we prepare each operator.

  • And they kind of cue up a bunch of allocations.

  • And those are all baked into a single pass where we then

  • allocate a single block of memory and tensors

  • are just fused into that large contiguous block of memory.

  • We don't yet support control flow.

  • But I will be talking about that later in the talk.

  • It's something that we're thinking about and working on.

  • It's on the near horizon for actual shipping models.

  • So what about the operator set?

  • So we support float and quantized types

  • for most of our operators.

  • A lot of these are backed by hand-tuned, neon, and assembly

  • based kernels that are specifically

  • optimized for ARM devices.

  • Ruy is our newest GEMM backend for TensorFlow Lite.

  • And it was built from the ground up with mobile execution

  • in mind, a [INAUDIBLE] execution.

  • We support about 120 built-in operators right now.

  • You will probably realize that that's quite

  • a bit smaller than the set of TensorFlow ops, which

  • is probably into the thousands by now.

  • I'm not exactly sure.

  • So that can cause problems.

  • But I'll dig into some solutions we have on the table for that.

  • I already talked about some of the benefits

  • of these high level kernels having

  • fused activations and biases.

  • And then we have a way for you to kind of, at conversion time,

  • stub out custom operators that you would like.

  • Maybe we don't yet support them in TF Lite

  • or maybe it's a one off operator that's not yet

  • supported in TensorFlow.

  • And then you can plug-in your operator implementation

  • at runtime.

  • So the hardware acceleration interface,

  • we call them delegates.

  • This is basically an abstraction that

  • allows you to plug in and accelerate

  • subgraphs of the overall graph.

  • We have NNAPI, GPU, EdgeTPU, and DSP backends on Android.

  • And then on iOS, we have a metal delegate backend.

  • And I'll be digging into some of these and their details

  • here in a few slides.

  • OK.

  • So what can I do with it?

  • Well, I mean this is largely a lot of the same things

  • that you can do with TensorFlow.

  • There's a lot of speech and vision-related use cases.

  • I think often we think of mobile inference

  • as being image classification and speech recognition.

  • But there are quite a few other use

  • cases that are being used now and are in deployment.

  • We're being used broadly across a number of both first party

  • and third party apps.

  • OK.

  • So let's start with models.

  • We have a number of models in this model repo

  • that we host online.

  • You can use models that have already

  • been authored in TensorFlow and feed those into the converter.

  • We have a number of tools and tutorials

  • on how you can apply transfer learning to your models

  • to make them more specific to your use case,

  • or you can author models from scratch

  • and then feed that into the conversion pipeline.

  • So let's dig into conversion and what that actually looks like.

  • Well, here's a brief snippet of how

  • you would take a saved model, feed that into our converter,

  • and output a TFLite model.

  • It looks really simple.

  • In practice, we would like to say

  • that this always just works.

  • That's sadly not yet a reality.

  • There's a number of failure points that people run into.

  • I've already highlighted this mismatch

  • in terms of supported operators.

  • And that's a big pain point.

  • And we have some things in the pipeline to address that.

  • There's also different semantics in TensorFlow that aren't yet

  • natively supported in TFLite, things like control-flow, which

  • we're working on, things like assets,

  • hash tables, TensorLists, those kinds of concepts.

  • Again, they're not yet natively supported in TensorFlow Lite.

  • And then certain types we just don't support.

  • They haven't been prioritized in TensorFlow Lite.

  • You know, double execution, or bfloat16, none of those,

  • or even FP16 kernels are not natively supported

  • by the TFLite built-in operators.

  • So how can we fix that?

  • Well, a number of months ago, we started a project called--

  • well, the name is a little awkward.

  • It's using select TensorFlow operators in TensorFlow Lite.

  • And effectively, what this does is

  • it allows you to, as a last resort,

  • convert your model for the set of operators

  • that we don't yet support.

  • And then at runtime, you could plug-in this TensorFlow

  • select piece of code.

  • And it would let you run these TensorFlow kernels

  • within the TFLite runtime at the expense of a modest increase

  • in your binary size.

  • What does that actually mean?

  • So the converter basically, it recognizes these TensorFlow

  • operators.

  • And if you say, I want to use them,

  • if there's no TFLite built-in counterpart,

  • then it will take that node def.

  • It'll bake it in to the TFLite custom operator that's output.

  • And then at runtime, we have a delegate

  • which resolves this custom operator

  • and then does some data marshaling

  • into the eager execution of TensorFlow, which again would

  • be built into the TFLite runtime and then marshaling

  • that data back out into the TFLite tensors.

  • There's some more information that I've linked to here.

  • And the way you can actually take advantage of this, here's

  • our original Python conversion script.

  • You drop in this line basically saying

  • the target ops set includes these select TensorFlow ops.

  • So that's one thing that can improve the conversion

  • and runtime experience for models that aren't yet

  • natively supported.

  • Another issue that we've had historically--

  • our converter was called TOKO.

  • And its roots were in this TFMini project,

  • which was trying to statically compute and bake

  • this graph into your runtime.

  • And it was OK for it to fail because it would all

  • be happening at build time.

  • But what we saw is that that led to a lot

  • of hard to decipher opaque error messages and crashes.

  • And we've since set out to build a new converter based

  • on MLIR, which is just basically tooling

  • that's feeding into this converter

  • helping us map from the TensorFlow dialect of operators

  • to a TensorFlow Lite dialect of operators

  • with far more formal mechanisms for translating

  • between the two.

  • And this, we think will give us far better debugging,

  • and error messages, and hints on how

  • we can actually fix conversion.

  • And the other reason that motivated this switch

  • to a new converter was to support control flow.

  • This will initially start by supporting functional control

  • flow forms, so if and while conditionals.

  • We're still considering how we can potentially

  • map legacy control flow forms to these new variants.

  • But this is where we're going to start.

  • And so far, we see that this will unlock

  • a pretty large class of useful models,

  • the RNN class type models that so far

  • have been very difficult to convert to TensorFlow Lite.

  • TensorFlow 2.0.

  • It's supported.

  • There's not a whole lot that changes on the conversion

  • end and certainly nothing that changes on the TFLite end

  • except for maybe the change to that saved model

  • is now the primary serialization format with TensorFlow.

  • And we've also made a few tweaks and added some sugar

  • for our conversion APIs when using quantization.

  • OK.

  • So you've converted your model.

  • How do you run it?

  • Here's an example of our API usage in Java.

  • You basically create your input buffer, your output buffer.

  • It doesn't necessarily need to be a byte buffer.

  • It could be a single or multidimensional array.

  • You create your interpreter.

  • You feed it your TFLite model.

  • There are some options that you can give it.

  • And we'll get to those later.

  • And then you run inference.

  • And that's about it.

  • We have different bindings for different platforms.

  • Our first class bindings are Python, C++, and Java.

  • We also have a set of experimental bindings

  • that we're working on or in various states of both use

  • and stability.

  • But soon we plan to have our Objective C and Swift

  • bindings be stable.

  • And they'll be available as the normal deployment libraries

  • that you would get on iOS via CocoaPods.

  • And then for Android, you can use our JCenter or BingeRate

  • ARs for Java.

  • But those are primarily focused on third party developers.

  • There are other ways you can actually reduce

  • the binary size of TFLite.

  • I mentioned that the core runtime is 100 kilobytes.

  • There's about 800 or 900 kilobytes

  • for the full set of operators.

  • But there are ways that you can basically

  • trim that down and only include the operators that you use.

  • And everything else gets stripped by the linker.

  • We expose a few build rules that help with this.

  • You feed it your TFLite model.

  • It'll parse that in output.

  • Basically, a CC file, which does the actual op registration.

  • And then you can rely on your linker

  • to strip the unused kernels.

  • OK.

  • So you've got your model converted.

  • It's up and running.

  • How do you make it run fast?

  • So we have a number of tools to help with this.

  • We have a number of backends that I talked about already.

  • And I'll be digging into a few of these

  • to highlight how they can help and how you can use them.

  • So we have a benchmarking tool.

  • It allows you to identify bottlenecks when actually

  • deploying your model on a given device.

  • It can output profiles for which operator's

  • taking the most amount of time.

  • It lets you plug in different backends

  • and explore how this actually affects inference latency.

  • Here's an example of how you would

  • build this benchmark tool.

  • You would push it to your device.

  • You would then run it.

  • You can give it different configuration options.

  • And we have some helper scripts that kind of help

  • do this all atomically for you.

  • What does the output look like?

  • Well, here you can get a breakdown of timing

  • for each operator in your execution plan.

  • You can isolate bottlenecks here.

  • And then you get a nice summary of where time is actually

  • being spent.

  • AUDIENCE: In the information, there

  • is just about operation type or we also

  • know if it's the earlier convolution of the network

  • or the later convolutions in the network or something like that?

  • JARED DUKE: Yeah.

  • So there's two breakdowns.

  • One is the run order which actually

  • is every single operator in sequence.

  • And then there's the summary where

  • it coalesces each operator into a single class.

  • And you get a nice summary there.

  • So this is useful for, one, identifying bottlenecks.

  • If you have control over a graph and then the authoring side

  • of things, then you can maybe tailor

  • the topology of your graph.

  • But otherwise, you can file a bug on the TFLite team.

  • And we can investigate these bottlenecks

  • and identify where there's room for improvement.

  • But it also affords--

  • it affords you, I guess, the chance

  • to explore some of the more advanced performance techniques

  • like using these hardware accelerators.

  • I talked about delegates.

  • The real power, I think, of delegates

  • is that it's a nice way to holistically optimize

  • your graph for a given backend.

  • That is you're not just delegating each op one by one

  • to this hardware accelerator.

  • But you can take an entire subgraph of your graph

  • and run that on an accelerator.

  • And that's particularly advantageous for things

  • like GPUs or neural accelerators where

  • you want to do as much computation on the device as

  • possible with no CPU interop in between.

  • So NNAPI is the abstraction in Android for accelerating ML.

  • And it was actually developed fairly closely in tandem

  • with TFLite.

  • You'll see a lot of similarities into the high level op

  • definitions that are found in NNAPI

  • and those found in TFLite.

  • And this is effectively an abstraction layer

  • at the platform level that we can hook into on the TensorFlow

  • Lite side.

  • And then vendors can plug in their particular drivers

  • for DSP, for GPUs.

  • And with Android Q, it's really getting to a nice stable state

  • where it's approaching parity in terms of features and ops

  • with TensorFlow Lite.

  • And there are increasingly--

  • there's increased adoption both in terms of user base

  • but also in terms of hardware vendors

  • that are contributing to these drivers more recently

  • we've released our GPO back end and we've also open source.

  • This can yield a pretty substantial speedup

  • on many floating point convolution models,

  • particularly larger models.

  • There is a small binary size cost that you have to pay.

  • But if it's a good match for your model,

  • then it can be a huge win.

  • And this is-- we found a number of clients

  • that are deploying this with things like face detection

  • and segmentation.

  • AUDIENCE: Because if you're on top of [INAUDIBLE] GPU.

  • JARED DUKE: Yeah, so on Android, there's a GLES back end.

  • There's also an OpenCL back end that's

  • in development that will afford a kind of 2

  • to 3x speed up over the GLES back end.

  • There's also a Vulcan back end, and then

  • on iOS, it's metal-based.

  • There's other delegates and accelerators

  • that are in various states of development.

  • One is for the Edge TPU project, which can either

  • use kind of runtime on device compilation,

  • or you can use or take advantage of the conversion step

  • to bake the compiled model into the TFLite graph itself.

  • We also announced, at Google I/O,

  • support for Qualcomm's Hexagon DSPs

  • that we'll be releasing publicly soon-ish.

  • And then there's some more kind of exotic optimizations

  • that we're making for the floating point CPU back end.

  • So how do you take advantage of some of these back ends?

  • Well, here is kind of our standard usage

  • of the Java APIs for inference.

  • If you want to use NnAPI, you create your NnAPI delegate.

  • You feed it into your model options, and away you go.

  • And it's quite similar for using the GPU back end.

  • There are some more sophisticated and advanced

  • techniques for both an API and GPU on our interop.

  • This is one example where you can basically use a GL texture

  • as the input to your graph.

  • That way, you avoid needing to copy--

  • marshal data back and forth from CPU to GPU.

  • What are some other things we've been working on?

  • Well, the default out of the box performance

  • is something that's critical.

  • And we recently landed a pretty substantial speed up there

  • with this ruy library.

  • Historically, we've used what's called gemmlowp

  • for quantized matrix multiplication,

  • and then eigen for floating point multiplication.

  • Ruy was built from the ground up basically

  • to [INAUDIBLE] throughput much sooner

  • in terms of the size of the inputs to, say, a given matrix

  • multiplication operator, whereas more

  • desktop and cloud-oriented matrix multiplication libraries

  • are focused on peak performance with larger sizes.

  • And we found this, for a large class of convolution models,

  • is providing at least a 10% speed-up.

  • But then on kind of our multi-threaded floating point

  • models, we see two to three times the speed-up,

  • and then the same on more recent hardware that has these neon

  • dot product intrinsics.

  • There's some more optimizations in the pipeline.

  • We're also looking at different types--

  • Sparse, fp16 tensors to take advantage of mobile hardware,

  • and we'll be announcing related tooling

  • and features support soon-ish.

  • OK, so a number of best practices here

  • to get the best performance possible--

  • just pick the right model.

  • We find a lot of developers come to us with inception,

  • and it's hundreds of megabytes.

  • And it takes seconds to run inference,

  • when they can get just as good accuracy,

  • sometimes even better, with an equivalent MobileNet model.

  • So that's a really important consideration.

  • We have tools to improve benchmarking and profiling.

  • Take advantage of quantization where possible.

  • I'm going to dig into this in a little bit

  • how you can actually use quantization.

  • And it's really a topic for itself,

  • and there will be, I think, a follow-up

  • session about quantization.

  • But it's a cheap way of reducing the size of your model

  • and making it run faster out of the box on CPU.

  • Take advantage of accelerators, and then

  • for some of these accelerators, you can also

  • take advantage of zero copy.

  • So with this kind of library of accelerators

  • and many different permutations of quantized or floating point

  • models, it can be quite daunting for many developers, probably

  • most developers, to figure out how

  • best to optimize their model and get the best performance.

  • So we're thinking of some more and working on some projects

  • to make this easy.

  • One is just accelerator whitelisting.

  • When is it better to use, say, a GPU or NnAPI versus the CPU,

  • both local tooling to identify that for, say,

  • a device you plugged into your dev machine

  • or potentially as a service, where we can farm this out

  • across a large bank of devices and automatically determine

  • this.

  • There's also cases where you may want to run parts of your graph

  • on different accelerators.

  • Maybe parts of it map better to a GPU or a DSP.

  • And then there's also the issue of when different apps are

  • running ML simultaneously, so you have a hotware detection

  • running at the same time you're running selfie segmentation

  • with a camera feed.

  • And they're both trying to access the same accelerator.

  • How can you coordinate efforts to make sure everyone's playing

  • nicely?

  • So these are things we're working out.

  • We plan on releasing tooling that can improve this

  • over the next quarter or two.

  • So we talked about quantization.

  • There are a number of tools available now

  • to make this possible.

  • There are a number of things being worked on.

  • In fact, yesterday, we just announced

  • our new post-training quantization

  • that does full quantization.

  • I'll be talking about that more here

  • in the next couple of slides.

  • Actually, going back a bit, we've

  • long had what's called our legacy quantized training

  • path, where you would instrument your graph at authoring time

  • with these fake quant nodes.

  • And then you could use that to actually generate

  • a fully quantized model as the output from the TFLite

  • conversion process.

  • And this worked quite well, but it was--

  • it can be quite painful to use and quite tedious.

  • And we've been working on tooling

  • to make that a lot easier to get the same performance

  • both in terms of model size reduction

  • and runtime acceleration speed-up.

  • AUDIENCE: Is part about the accuracy-- it seems

  • like training time [INAUDIBLE].

  • JARED DUKE: Yeah, you generally do.

  • So we first introduced this post-training quantization

  • path, which is hybrid, where we are effectively just quantizing

  • the waits, and then dequantizing that at runtime

  • and running everything in fp32, and there

  • was an accuracy hit here.

  • It depends on the model, how bad that is,

  • but sometimes it was far enough off

  • the mark from quantization aware training

  • that it was not usable.

  • And so that's where--

  • so again, with the hybrid quantization,

  • there's a number of benefits.

  • I'm flying through slides just in the interest of time.

  • The way to enable that post-training quantization--

  • you just add a flag to the conversion paths,

  • and that's it.

  • But on the accuracy side, that's where we came up

  • with some new tooling.

  • We're calling it per axis or per channel quantization, where

  • with the waits, you wouldn't just

  • have a single quantized set of parameters

  • for the entire tensor, but it would

  • be per channel in the tensor.

  • And we found that that, in combination with feeding it

  • kind of an evaluation data set during conversion time,

  • where you would explore the range of possible quantization

  • parameters, we could get accuracy that's almost on par

  • with quantization aware training.

  • AUDIENCE: I'm curious, are some of these techniques

  • also going to be used for TensorFlow JS,

  • or did they not have this--

  • do they not have similarities?

  • They use MobileNet, right, for a browser?

  • JARED DUKE: They do.

  • These aren't yet, as far as I'm aware, used or hooked

  • into the TFJS pipeline.

  • There's no reason it couldn't be.

  • I think part of the problem is just very different tool

  • chains for development.

  • But--

  • AUDIENCE: How do you do quantized operations

  • in JavaScript?

  • [INAUDIBLE]

  • JARED DUKE: Yeah, I mean I think the benefit isn't

  • as clear, probably not as much as if you were just

  • quantizing to fp16.

  • That's where you'd probably get the biggest win for TFJS.

  • In fact, I left it out of these slides,

  • but we are actively working on fp16 quantization.

  • You can reduce the size of your model by half,

  • and then it maps really well to GPU hardware.

  • But I think one thing that we want to have is

  • that quantization is not just a TFLite thing,

  • but it's kind of a universally shared concept

  • in the TensorFlow ecosystem.

  • And how can we take the tools that we already

  • have that are sort of coupled to TFLite

  • and make them more generally accessible?

  • So to use this new post-training quantization

  • path, where you can get comparable accuracy to training

  • time quantization, effectively, the only difference

  • here is feeding in this representative data

  • set of what the inputs would look like to your graph.

  • It can be a-- for an image-based model,

  • maybe you feed it 30 images.

  • And then it is able to explore the space of quantization

  • and output values that would largely

  • match or be close to what you would

  • get with training-aware quantization.

  • We have lots of documentation available.

  • We have a model repo that we're going to be investing

  • heavily in to expand this.

  • What we find is that a lot of TensorFlow developers--

  • or not even TensorFlow developers-- app developers

  • will find some random graph when they search Google or GitHub.

  • And they try to convert it, and it fails.

  • And a lot of times, either we have

  • a model that's already been converted

  • or a similar model that's better suited for mobile.

  • And we would rather have a very robust repository

  • that people start with, and then only

  • if they can't find an equivalent model,

  • they resort to our conversion tools or even authoring tools.

  • AUDIENCE: Is there a TFLite compatible section in TFHub?

  • JARED DUKE: Yeah, we're working on that.

  • Talked about the model repo training.

  • So what if you want to do training on device?

  • That is a thing.

  • We have an entire project called The [INAUDIBLE]

  • Team Federated Learning Team, who's focused on this.

  • But we haven't supported this in TensorFlow Lite

  • for a number of reasons, but it's something

  • that we're working on.

  • There's quite a few bits and components that still have yet

  • to land to support this, but it's something

  • that we're thinking about, and there is increasing demand

  • for this kind of on-device tuning or transfer

  • learning scenario.

  • In fact, this is something that was announced at WWDC, so.

  • So we have a roadmap up.

  • It's now something that we publish publicly

  • to make it clear what we're working on,

  • what our priorities are.

  • I touched on a lot of the things that

  • are in the pipeline, things like control flow and training,

  • proving our runtime.

  • Another thing that we want to make easier

  • is just to use TFLite in the kind of native types

  • that you are used to using.

  • If you're an Android developer, say, if you have a bitmap,

  • you don't want to convert it to a byte buffer.

  • You just want to feed us your bitmap, and things just work.

  • So that's something that we're working on.

  • A few more links here to authoring apps

  • with TFLite, different roadmaps for performance and model

  • optimization.

  • That's it.

  • So any questions, any areas you'd

  • like to dive into more deeply?

  • AUDIENCE: So this [INAUDIBLE].

  • So what is [INAUDIBLE] has more impact like a fully connected

  • [INAUDIBLE]?

  • JARED DUKE: Sorry.

  • What's--

  • AUDIENCE: For a speed-up.

  • JARED DUKE: Oh.

  • Why does it?

  • AUDIENCE: Yeah.

  • JARED DUKE: So certain operators have been,

  • I guess, more optimized to take advantage of quantization

  • than others.

  • And so in the hybrid quantization path,

  • we're not always doing computation in eight-bit types.

  • We're doing it in a mix of floating point and eight-bit

  • types, and that's why there's not always the same speed-up

  • you would get with like an LSTM and an RNN versus a [INAUDIBLE]

  • operator.

  • AUDIENCE: So you mentioned that TFLite is

  • on billions of mobile devices.

  • How many apps have you seen added to the Play Store that

  • have TFLite in them?

  • JARED DUKE: Tim would have the latest numbers.

  • It's-- I want to say it's into the tens of thousands,

  • but I don't know that I can say that.

  • It's certainly in the several thousands,

  • but we've seen pretty dramatic uptick, though, just

  • tracking Play Store analytics.

  • AUDIENCE: And in the near term, are you thinking more

  • about trying to increase the number of devices that

  • are using TFLite or trying to increase

  • the number of developers that are including it

  • in the applications that they built.

  • JARED DUKE: I think both.

  • I mean there are projects like the TF Micro, where

  • we want to support actual microcontrollers

  • and running TFLite on extremely restricted, low power arm

  • devices.

  • So that's one class of efforts on--

  • we have seen demand for actually running TFLite in the cloud.

  • There's a number of benefits with TFLite

  • like the startup time, the lower memory footprint that

  • do make it attractive.

  • And some developers actually want

  • to run the same model they're running on device in the cloud,

  • and so there is demand for having

  • like a proper x86 optimized back end.

  • But at the same time, I think one of our big focuses

  • is just making it easier to use-- meet developers

  • where they're at.

  • And part of that is a focus on creating

  • a very robust model repository and more idiomatic

  • APIs they can use on Android or iOS

  • and use the types they're familiar with,

  • and then just making conversion easy.

  • Right now, if you do take kind of a random model

  • that you found off the cloud and try

  • to feed it into our converter, chances

  • are that it will probably fail.

  • And some of that is just teaching developers

  • how to convert just the part of the graph they want,

  • not necessarily all of the training that's surrounding it.

  • And part of it is just adding the features and types

  • to TFLite that would match the semantics of TensorFlow.

  • I mean, I will say that in the long run,

  • we want to move toward a more unified path with TensorFlow

  • and not live in somewhat disjoint worlds,

  • where we can take advantage of the same core runtime

  • libraries, the same core conversion pipelines,

  • and optimization pipelines.

  • So that's things that we're thinking about for the longer

  • term future.

  • AUDIENCE: Yeah, and also [INAUDIBLE]

  • like the longer term.

  • I'm wondering what's the implication of the ever

  • increasing network speed on the [INAUDIBLE] TFLite?

  • [INAUDIBLE],, which maybe [INAUDIBLE] faster than current

  • that we've [INAUDIBLE] take [INAUDIBLE] of this.

  • JARED DUKE: We haven't thought a whole lot about that,

  • to be honest.

  • I mean, I think we're still kind of betting on the reality

  • that there will always be a need for on device ML.

  • I do think, though, that 5G probably

  • unlocks some interesting hybrid scenarios, where you're

  • doing some on device, some cloud-based ML,

  • and I think for a while, the fusion of on device hotware

  • detection, as soon as the OK, Google is detected,

  • then it starts feeding things into the cloud.

  • That's kind of an example of where there is room

  • for these hybrid solutions.

  • And maybe those will become more and more practical.

  • Everyone is going to run to your desk

  • and start using TensorFlow Lite after this?

  • AUDIENCE: You probably already are, right?

  • [INAUDIBLE] if you have one of the however many apps that was

  • listed on Tim's slide, right?

  • JARED DUKE: I mean, yeah.

  • If you've ever done, OK, Google, then you're

  • using TensorFlow Lite.

  • AUDIENCE: [INAUDIBLE].

  • Thank you.

  • JARED DUKE: Thank you.

  • [APPLAUSE]

JARED DUKE: Thanks everybody for showing up.

字幕と単語

ワンタップで英和辞典検索 単語をクリックすると、意味が表示されます

B1 中級

TensorFlowの内部:TensorFlow Lite (Inside TensorFlow: TensorFlow Lite)

  • 2 0
    林宜悉 に公開 2021 年 01 月 14 日
動画の中の単語