字幕表 動画を再生する 英語字幕をプリント [MUSIC PLAYING] CHET HAASE: Hey. [CLAPPING] DAN SANDLER: Hey everybody. Welcome back to What's New in Android. I'm Dan Sandler from the System UI team. CHET HAASE: I'm Chet Haase from the Android Toolkit team. ROMAIN GUY: And I'm Romain Guy from-- I still don't have a name for my team. So from the Android team. [LAUGHTER] And you may remember us from other talks, such as What's New in Android 2017, What's New in Android 2016, What's New in Android 2015, What's New in Android 2014. Not What's New in Android 2013. CHET HAASE: We don't talk about that one. ROMAIN GUY: We don't talk about that one. DAN SANDLER: You know, that was the time that we had Jelly Bean two years in a row. It was brilliant. We didn't have to redo the logo or anything. [LAUGHTER] ROMAIN GUY: But now-- CHET HAASE: What's New in Android, what I like to call the Android keynote. Nobody else does. But I like to call it that. Because this is where we talk to you about all of the developer stuff going on in the Android platform. In particular, let's talk about Android P. Specifically, let's talk today about-- oh, hang on. Android APIs. [LAUGHTER] ROMAIN GUY: All right, first we'll start with distribution. You saw in the keynote we introduced the dynamic ad bundles. [? Tor's ?] demo was pretty clear. It's pretty easy for you. All you have to do is click your different menu when you build your application. And we're going to save you some space. It's going to be faster and easier for your users to download your app. And I'm sure you have a lot of questions about it. So we have a couple of talks this afternoon, at 5:00 PM and 6:00 PM. Go there if you want answers, because we don't have them. CHET HAASE: So you're going to see a slide a whole lot like this through the rest of the talk. I feel like the main function we serve in this talk is to tell you the other talks to go to. We're like the appendix. We're like the index for the rest of the content. DAN SANDLER: I mean, we're like obsolete, like vestigial. Is that what it is? CHET HAASE: I don't like to think about that. ROMAIN GUY: Yeah. Let's be clear. A lot of people back at work have done all the hard work. We just get to go on stage and talk about their hard work. CHET HAASE: So speaking of that, let's talk about Android Jetpack. We heard Steph talk about this in the developer keynote. This is a set of components, as well as guidance on how to build better Android applications. All of you are familiar with most of what is in Android Jetpack already. What we're doing is adding to it over time with stuff that's going to make it even better. And we're also improving it over time. One of the major steps that we're taking is what I like to think of as a refactor, because it's a refactor. My favorite thing about the support library is how the package names embed the release number in them. So, for example, we support things like v4. Actually we don't support v4 anymore. We have a min SDK of at least 14 now. But it's in the package name. Isn't that a little bit silly? So we're doing away with that. We're doing a whole lot of tedious renaming. And we're also providing tools to make it easier for you to do the similar refactoring that you're going to need to do in your application, as well as in Android Studio. Everything is being renamed to something more appropriate called Androidx. If you want to know more about the details of that, the renaming, as well as doing more modular, more fine-grained splits to make sure that you don't drag in too much stuff, go to the talk. Learn what's new in Android Support Library. Also, there was an article that was posted on the Android Developers blog about a half hour ago. Check that out for more of the details there. Let's talk about Android Test, which is part of this new Jetpack thing going on. Android Test is the ATSL stuff, the Espresso stuff that hopefully you were already using really good ways to test your application. Now, they provide first class Kotlin support, as well as more elegant APIs for reducing a lot of the boilerplate. Here's a simple example. We used to have a way of asserting, which, A, was not necessarily obvious in the parameters you were passing. Also, unobvious in the order of the parameters that you were passing. And then it would give you an error message that also didn't really help very much. So we have something a little more sensible now. You can assert that it's actually working on the visible property. And the error message gives you something more that you can work with a little bit better. Go to the Frictionless Android Testing talk for more information about that stuff. Jetpack architecture is about the architecture components that were announced last year at I/O and then iterated with feedback from the community, and finally went 1.0 in the fall. So we have the release parts of those, which includes all the Lifecycle stuff, and the ViewModel stuff, as well as the Room, the persistent data model stuff, and LiveData. So hopefully you are using that stuff already, at least in your new applications. And what we also have is recently we released the paging library for doing asynchronous data paging into RecyclerView. That was alpha, then beta, because that's how those things work. And now it's 1.0 this week. So please start using that. And we also talked in the developer keynote about a couple of new things that you should check out soon. WorkManager is currently in preview. There's going to be a talk about it. It's about job scheduling, but job scheduling in a way where we handle all the cases back in previous releases, instead of you having to use specific approaches, depending on what version and device that you're on. Also, navigation. It turns out that up versus back is a hard problem for applications to solve. We are making that much easier. And we're integrating with the tool to make it even easier yet. So go to all these talks. There's an overview talk, as well as specific talks on Navigation Controller and WorkManager, and also a talk on RecyclerView and Paging. Me? Again? ROMAIN GUY: It says your name on the slide. CHET HAASE: I keep building suspense into this thing. What's going to happen next? Who's he going to hand the clicker to? It's still mine. DAN SANDLER: It's still you. CHET HAASE: OK. Let's talk about battery. This is one of the ongoing efforts in Android to help the users, because it turns out battery is really important. We're all power users. Unfortunately, we just keep using the power. So what can we do about it? We can create these app standby buckets. We're going to monitor the usage of applications, and see how actively the user is using it, and then make determinations about how much access that application has to ongoing things in the system which are going to take up more battery. We also had background restrictions that the user has the ability to kick in in settings. So if an application is behaving badly, let's say holding wake locks for long periods of time, or waking up constantly, or accessing services way more than they should when it's not on a charger, then we'll note that and expose that in settings. And then the user can take some action on that if they deem that necessary. Go to the battery session on Thursday morning to learn more of the details there. ROMAIN GUY: So one of the things that we've been focusing on with Android P is privacy. Maybe that's what it stands for. So one of the things we've done is that when your app is in the background, it doesn't have access to the microphone anymore. It doesn't have access to the camera anymore. And it doesn't have access to the sensor kind of. So you won't receive the data from the sensors automatically. You can manually pull from the sensors, and you'll get batch updates, but the best thing to do if you want to get access to the sensor data is to keep a foreground service running instead. So no more microphone. No more camera for you. I think I've heard in the past that some apps were trying to stay alive in memory by playing a white noise or keeping the microphone on. Don't do that anymore. It's not OK. Kotlin-- it's this little thing we announced last year. So we are busy. We want to make it better for all the Kotlin developers out there. I'm sure there's a lot of you here today. So some of the things we've been doing, the R team has been really busy with D8, R8, and ART itself. They've been looking at the byte code generated by the Kotlin compiler. They've analyzed some of the byte code patterns that were different from the ones generated by the Java programming language compiler. And they've been optimizing for those patterns. We've also been adding a lot of nullability annotations to our Java APIs, both in the core libraries, so libcore, and our support libraries to make it easier for you to use the platform APIs when you're in Kotlin. And finally, we launched on GitHub a new library called Android KTX. It's a set of Kotlin extensions for existing platform APIs. And the goal here is to try to take advantage of some of the Kotlin language features to make existing APIs easier to use. They're already easier to use just by using Kotlin, but with the extension they get even better. And I want to thank the community, because we've received dozens of pull requests, and also bugs and feature requests from you. And we've accepted a bunch of them. So if you have ideas, if you have things that you would like to see in Android KTX, please go to GitHub, and we'll take a look at your PR. And this is an example of the kind of code you can write with KTX. If you want to create a bitmap, you don't have to specify its ARGB 8.8.8.8 anymore. You can call applyCanvas, which automatically creates the canvas for you. It becomes this. And at the bottom you can see, for instance, the destructuring assignment for color integer. So you don't have to do any shifting or masking of the int into bytes. We'll take care of that for you. There is a talk by Jack Wharton on Thursday at 10:30 AM. He's going to go through most of the extensions. He's going to talk about the philosophy behind the extensions, how we write them, what are the kind of extensions that we want to see in that library that we're not looking for in that library. So before you do all the work and send a PR, go attend that talk to understand what we're looking for. CHET HAASE: We already talked about the Android Test stuff that is part of Jetpack earlier. That's probably a better, more holistic way to test your application. But if there is a specific situation in which you find it necessary or helpful to mock the framework-- and I don't mean ridicule it, because that wouldn't be nice-- then it is possible to do that in easier ways now in Mockito. We are not changing the framework, but we're actually integrating changes into Mockito itself. So you can now mock final methods. And soon you should be able to mock static methods. [CLAPPING] [CHEERING] ROMAIN GUY: And Chet is making that face because he doesn't understand why that is so interesting to you. [LAUGHTER] CHET HAASE: And system created objects like activity, eventually. Working on that internally, but it should be on the way eventually. A background text measurement. This is part of a bunch of smaller changes that we made in the text area. It turns out that measurement is really expensive. So most applications do text. And I would bet that the text operations in your application are some of the most expensive going on in the UI thread, which can contribute to jank. Wouldn't it be nicer if you could offload that to a background thread so that by the time you actually need to render the text, or perform those operations on the UI thread, most of the hard work was done for you? So the observation is that 80% to 90% of the operations that are necessary for actually displaying text happen in text measurement. Well, we've made it possible and much easier to actually perform this as a background operation. So we have a class called pre-computed text. And you can query that. And you can say, I want you to pre-measure this. And then you can set that text that's spannable on the text view later whenever you want. So you do it in background threads, like this. You say create this thing. And then you can set that on the text view later when you actually need it. Should be much faster. Magnifier is something that if you're using the preview releases you might see, if you select some text, it pops up this little bar up above it. It makes it easier to manipulate the cursor. That's really great for text, but the other cool thing about it is that it's also available for your applications for any other use case. So there's an API that allows you to pop up this magnifier for whatever happens to be in your view. So you can show this, and dismiss it, and use it for your stuff as well. So core functionality that we want in the system user interface. But also useful APIs for developers to use for their own purposes. I don't know if you've worked with your design department, and they've specified something about text in your UI, and, OK, well, I want this aligned this many dips away from the top. And I want the baseline on the bottom this many dips away from the bottom. And then you have this interspersed vertical alignment stuff going on. And then you sort of puzzle with this for a while. And you basically futz with padding in all kinds of different configurations to sort of get it to where they wanted to get it. We have worked with that and created some new attributes, and methods, and properties for you to use that make that much easier. So we allow you to just pass us the information about the baseline alignment calculations that you would like to perform. And then we'll futz with padding on your behalf. Smart Linkify. I think of this as being Linkify, but smarter. So we already have the ability to ask for links in a block of text, and it'll detect things like phone numbers and addresses. But we also have the ability, through machine learning models, and stuff that you've seen through smart text selection, to detect other entities. We can do the same thing with Linkify. It takes a little bit potentially longer. So you do this off thread. So you would basically generate the links off thread, and then set it on your text view later, using code similar to this. There's a text talk on Wednesday evening. So please go to that for more details about all of this, as well as more. ROMAIN GUY: So location. Now you can take advantage of a new package, android.net.wifi.rtt. It's the WiFi round trip time API. It requires compatible hardware on your phone. It also requires a compatible access point. And it allows you to find a precise location indoors for the user's device. You need to request the Find Location permission. And you don't need to connect to the access point. So if you're building an application that requires locating the user inside a big building, you can take advantage of this API in Android P. CHET HAASE: Accessibility has some improvements for navigation through the app. So it's easier for you to declare these functional blocks. It makes it easier for accessible users to understand how things are being grouped on the screen. There's an important talk on accessibility right now? No, that's tomorrow. OK. Hopefully it's tomorrow. I'm not sure that's correct. ROMAIN GUY: It's tomorrow or 40 minutes ago. CHET HAASE: OK. All right. DAN SANDLER: It's on YouTube. CHET HAASE: Yes, it is. Eventually. Like now. Yeah. I think I got the day wrong. Sorry about that. It's now. If you're in the wrong talk, I invite you to go to the other one. [LAUGHTER] ROMAIN GUY: Oh, my turn. Security. New API in Android P, the Unified BiometricDialog. So we deprecated the FingerprintManager because there are more ways to authenticate yourself with your body than just with a fingerprint. Could be your eyes. Could be whatever else that device manufacturers will think of next. So now we have a single UI for all devices and all means of authentication. We also have stronger protections for private keys. And very important in your application, if you're using the API, Build.SERIAL, it doesn't work anymore. The API is still there. But it basically returns bogus data. So you cannot rely on it at all anymore. CHET HAASE: Various changes in Enterprise. Just a couple of them that are interesting. We made it easier to work with work profile apps or different profile apps by having these different tabs that you can associate with them so they're not all mixed together, but you can actually have these whole sections of these different profiles. Also, you're allowed to lock packages to a specific task. You could have a launcher with just a minimum set of a group or a single application. That works in combination with the ability to have ephemeral users, which now gives you kiosk mode. So you will no longer have experiences like I had on a recent flight, as you can see from my blurry picture, where you see a movie and you wonder what operating system is running under that. So you swipe from the bottom of the screen. You see the ICS navigation bar. And then you press on the recent tasks, and you swipe the movie away, and you confuse the heck out of the system. ROMAIN GUY: I have to admit, that's what I try every time I'm on a plane these days. And it works a surprising number of times. DAN SANDLER: Homework for all of you. CHET HAASE: This is like fun for Android engineers. This is what we do. ROMAIN GUY: Hopefully this is not linked to the plane navigation. DAN SANDLER: Now? Is it OK? Can I talk now? Is that all right? CHET HAASE: Very briefly. DAN SANDLER: Very briefly. All right. Let's talk about actually a lot of the system UI stuff typically gets shown at one of the keynotes that precedes What's New in Android. So you've all seen a lot of the great stuff that users are getting. I'm going to talk to you about some of the stuff that as developers you might be interested in. The first one is display cutouts, a.k.a.-- well, there are other names for it. These are coming to the ecosystem all over the place. And so as a developer, you need to know where it's safe to draw and where it isn't. When you get your window insets on viewWindowInset, something or other, you get DisplayCutout object, which gives you all kinds of interesting data about the cutout. But you're probably going to want to use something called windowLayoutInDisplayCutoutMode on your window. So there's the basic way, which is, I never want to overlap the cutout. Just leave a black bar at the top or the bottom. Whatever. I'm not all that interested. A little more advanced would be display cutout mode default, which is, if you were already going to clear the status bar just fine, we'll let the app window draw underneath the cutout as well. So you'll get the nice action bar color extending through the status bar and extending underneath the cutout. Better still, or more advanced still is shortEdges cutout mode, which means essentially any time there's a cutout on the short edges of the device, and we're in portrait, I will just draw underneath. You don't have to do anything special. And in that situation you do need to look at the display cutout and ask it for the safe insets. Essentially, well, OK, I'm drawing everywhere. But you tell me what single rectangle of the screen is safest to draw in. And then finally, the cosmic version of this. You can do shortEdges, but you can actually get the bounds of the cutout as a region. So you will get the exact set of rectangles that are unavailable to you on the screen so you can display UI in the corners if the corners are available. Or if there's a corner cutout, you can move things out of the way so that it's visible in the center of the display. This is the most advanced version of it. And you can put the electrical tape away, because you can actually simulate notches now in developer options on your device. This is really, really exciting. Slices actually was what we were calling it internally. And we liked it so much we just kept it. You've seen Slices now in a couple of keynotes. It's essentially something that we've discovered and learned about on system UI and in the toolkit over many years of dealing with remote views for app widgets dealing with notifications. Essentially the problem of getting content from your app into some other place. So Slices is our new approach to remote content that you can actually use to project UI into your own app, or into other apps that support it. It's very structured. This is not sort of, here's a canvas, or an absolute layout. Go nuts with it. We give you a structure to fill out, and a whole bunch of very flexible templates in which to populate that data with some display hints so that the receiving end of the slice, the slice host, kind of knows what to do with it. These are interactive. These are updateable. This is meant to be something that holds rich UI. Sliders, controls, live information, possibly videos, things that actually feel like real UI, as opposed to a snapshot of something happening in a distant process somewhere. Slices are addressable by content URI. And this is how they're passed around the system, and how they're passed along to app indexing to be shown in context like search. And then finally, Slices is entirely inside the support library. It's entirely in Jetpack. So it's backwards compatible. You can use Slices all the way back to API 19. There's a great talk about Slices tomorrow bright and early, building interactive results for Google Search. Come to find out more about how all this technology works and how you can build your own. Related to Slices is Actions. You can think of these as shortcuts with parameters. Romain likes to think of them as visible intent. This is essentially a deep link into your app with some additional payload. It's not just a link to music. It's linked to the particular album, or something like that. And you saw these as well in the keynotes, showing up in a predictive space inside our app launching experience, actions you define in an actions XML file that goes into your APK or app bundle. And that too can get registered with app indexing so that search results and predictive features can show those actions. And there's a talk about this too, Thursday, slightly less early in the morning. Integrating your Android apps with the Google Assistant. Notifications. There was a lot of great stuff about digital wellness and controlling notifications that you saw in the keynote. And I'm very excited about it in P. I'm going to talk about some of the developer stuff, though, that we have in here. We asked users what notifications are most important to them. Users love messages. So we focused our energy in enhancing the messaging style API. You can do inline images now. You can do participant images, and attach other metadata about the participant. And we finally now have UI for smart reply, which we've had on Android Wear for years. So when you use RemoteInput.setChoices, those will now appear as chips right in the notification so you can respond instantly in the middle of a chat without leaving the notification shade. And there's tons of other stuff as usual. I had one other slide that we added about 10 minutes ago to this deck about notifications. And I'm just going to let that sit on the screen for a little while. [CLAPPING] So if you're doing something in the background, the user still needs to know. But Android P does a much better job of allowing notifications that you may already have running testify to that background activity, including things like overlaying windows. So with that, talk about the runtime, Romain. ROMAIN GUY: So probably one of the most important things for you as an Android developer is to understand our deprecation policy. It was announced a few weeks ago. Soon we will require all the applications to target some of the newest API levels. And we're doing that to make sure that we can keep the security level of Android as high as possible, as well as performance, and a lot of other nice things. So what does it mean for you? As of August this year, new applications published on the Play Store will have to target API level 26. And as of November this year, any update to an existing application that you publish on the Play Store will have to target API 26. And you can expect those numbers to go up over time. If you have native code in your application, we've been supporting 32-bit and 64-bit for a few years now. We will make 64-bit APIs required as of August of next year. You'll still be able to ship 32-bit support in your application, but we will ask you to ship 64-bit as well. One of the reasons to ship 64-bit on 64-bit devices is that you get better performance and much better code out of it. If you want to know more about the deprecation policy, there is a talk on Thursday afternoon. And I'm sure you'll have a lot of questions for the folks there. App compatibility. If you've tried the developer preview in P, you might have noticed something different if you're one of those naughty applications that uses some of our private APIs. On Android we have two types of private APIs, the APIs that are actually marked private. And then there's this weird @hide thing that we use in our source code. It's a Java Doc tag that we process especially to indicate that this is a public API for us. Not for you. Just for us. And we're a bit jealous because a lot of you are using them. So from now on, a lot of those APIs will trigger warnings in one form or another. It might be toast, or it might be logs when you make an illegal call. For some of these APIs we need to hear from you that you need those APIs for your application to keep working. Sometimes it's just an oversight. We didn't make the API public. It's just because we didn't think about it. So please go to this URL, and let us know if there's an API that you think should be made public. We might say yes. We might also say no. We have three types of lists. I won't go into too much detail here. But basically, if an API falls into the blacklist, you will not be able to call it ever. I'm not sure if we have anything in the blacklist right now. But those will evolve over time. So, again, let us know what APIs you need. CHET HAASE: This is one of the important reasons. This is why we ship previews. We need you to go out there and try your applications on the previews, because this is the time to find out when there's problems that either you can fix before the real release is out there, or you can let us know if it's a problem that we need to work on instead. ROMAIN GUY: And that's Chet trying to increase his engagement with his podcast. So there's this podcast called ADB, Android Developer Backstage. And in episode 89, they had-- who did you have in the episode? CHET HAASE: Brian Carlstrom. ROMAIN GUY: Brian Carlstrom. And they talked about this app compatibility issue, and what it means for you. So go listen to it. NDK. The release 17 of the NDK brings a lot of very interesting things. So, first of all, the neural networks APIs that were part of API level 27. We also have a new shared memory API if you do a lot of JNI. More importantly, we have finally ASAN, the address sanitizer to make sure your code doesn't scribble all over the memory. Now you don't need a rooted device anymore to use it. And we also have an undefined behavior sanitizer. It can be very difficult to detect undefined behaviors in your C or C++ code. So now there's a tool for that. We finally removed support for the deprecated APIs. So if you still use ARMv5 or MIPS, MIPS 32-bit or 64-bit, support is gone. You should not be shipping those anymore. In the upcoming release, release 18, we will remove GCC. So the GCC compiler was deprecated last year. Now everything is compiled with Clang in the NDK. We think we gave enough time. So GCC is going away. So if you're still using it, maybe you should not be in this talk. You should go fix your code. [LAUGHTER] And finally, we added support for the simpleperf CPU profiler. And we also have support in Android Studio for native profiling. So you don't even have to type anything in the terminal. Graphics and media. The camera APIs are getting better and better. We've added a lot of things that we were using ourselves in the camera applications. For instance, we give you access to the timestamps of the optical image stabilization. So if you want to build the kind of stabilization we've built in the video recording part of the camera application, now you can. If your app is doing a lot of selfies and using the display as a flash, you can tell the camera that you're doing that so you can adapt the exposure accordingly. We have support for USB cameras. I haven't seen any use for it, but I've heard some of you ask about it. So now it's available. Multi-camera support. There are some phones coming out out there with multiple cameras in the back, or in the front I suppose. And now we can expose them as a logical camera that contains more than one stream of data. ImageDecoder. I'm sure a lot of you are familiar with BitmapFactory. And if you're not fond of that API, trust me, we're not either. So there's a new one called ImageDecoder. It's part of Android P. The idea is to make it not only easier to decode images, but also to make it possible to decode animated images. So ImageDecoder can decode bitmaps, but also drawables, including animated drawables. CHET HAASE: There's this new thing called animated GIFs. I don't know. All the kids are using them. It's possible to actually load those now. ROMAIN GUY: So there are a few concepts that you have to learn when you learn ImageDecoder. And we're going to go through them in an example. So we have the concept of the source. We have the concept of the post processor, and finally, the header listener. So this is what the API looks like. First, you have to call createSource on the ImageDecoder. The source can be an asset. It could be a file descriptor. It can be many different things. And the goal here is that once you create a source, you can decode multiple images from the same source. This is particularly useful if you want to build thumbnails. You can decode the same source once at high resolution, once at lower resolution, or even intermediate resolutions. And that can be done from multiple worker threads. Then you can call decodeBitmap. You pass the source. And you can optionally pass a header listener. So the header listener, we have a lambda. It gives you back the decoder itself, the metadata about the image and the source. And it's inside the header listener that you can set the options. So in BitmapFactory you had bitmapfactory.options that you had to pass through Decode. Here you have to wait for the header listener to be invoked. CHET HAASE: We should point out too, that set target size, that's kind of a fundamental difference where the old bitmap-- DAN SANDLER: It's up there for them. CHET HAASE: No. I know. I see that. It's right there it. [INTERPOSING VOICES] CHET HAASE: That's a huge difference. Like before, if you wanted the right target size, you needed to work with, what? In density? As well as? ROMAIN GUY: In sample size, in density, and do a lot of trickery. CHET HAASE: We didn't know how they worked either. ROMAIN GUY: Yeah. [LAUGHTER] CHET HAASE: So a lot easier now. ROMAIN GUY: Yeah. Usually our answer was, figure it out. Anyways, so now you can just tell us what size you want the bitmap to be in. And we'll take care of it, finally. And you can also set the post processor. So the post processor is something with a simple interface. It gives you a canvas so you can draw on the bitmap right after it's decoded so you can add a header, like your title, or some metadata on it. Media. We are adding support in the platform for the HDR profile of VP9 so you can play HDR videos in your applications. YouTube was doing it, but they had their own way of decoding videos. If you own a device that doesn't support HDR, the playback will be in low dynamic range instead. But on a device that's capable, like a Pixel 2, you will be able to see the HDR stream in all its glory. We're also adding support, and bear with me, cause it's a little bit confusing, a format called HEIF. Yeah. It rolls off the tongue. The High Efficiency Image Format. So it's based on HEVC, which was also called H.265. And the filename extension is commonly HEIC. I don't know why they used different letters. But that's what they did. So we have to deal with it. So it is a container. And it can store multiple images. So you can use it to store a single image with higher quality and higher compression ratios than JPEG, for instance. JPEG is a lot easier to say. I like JPEG. But you can also sew multiple images if you want animated images or short movies. This is part of the support library, or, I guess, Jetpack now. And this is not part of the compressed API that you find on bitmap. Because it's a container, it works differently. So this is what it looks like. You have to create a builder. You have to tell us the path where you want to output the file. You have to give us in advance the width and the height of the image. And the source of the image can come from a bitmap. But it can also come from a surface. So you don't necessarily have, for instance, if you're doing GR rendering, or if you're doing video playback, you can include that directly as an image. You don't have to go through an intermediate bitmap. And then when you call writer to choose, you can add multiple bitmaps. In this case, we're adding only one. And when you call stop, we write it out. CHET HAASE: I should point out too that even though it's spelled HEIC, it's pronounced GIF. [LAUGHTER] All right. DAN SANDLER: Go. ROMAIN GUY: What are you still doing here? Vulkan. So this is the slide that I'm excited about. Very few of you will be. And I'm sure you will be super excited by things like subgroup ops and YCrCb formats. More seriously, anybody who's building middleware, for instance, the Unity engine, the Orion engine, or if you're building games, Vulkan is a low level graphics API that gives you a lot more control over the GPU so you get higher performance code. And Vulkan 1.1 adds new capabilities that were not possible in OpenGL. But it also closes the gap with OpenGL. So, for instance, the support of protected content. We were not able to play back protected content in Vulkan before. So that will unlock a lot of things, like video players, for instance. Neural networks APIs. They are not technically part of P. I mean, they are part of P, but we announced them in API level 27. You might have missed that. It's a C API that's designed for machine learning. It's also a fairly low level API. It's meant for basically playback of trained models. So things like TensorFlow are built on top of your neural network API. So you would use TensorFlow to do all the learning. And the neural network APIs can do basically the inference on device. The interesting thing is that we're also unlocking the access to the DSP that we use on Pixel 2 so you can get hardware accelerated machine learning effectively on select devices. ARCore you just saw in the developer keynote. A couple months ago we introduced ARCore 1.0. And they just announced ARCore 1.2. This takes care of tracking things in the real world. The problem is that you had to write a lot of OpenGL code. And it's not always pleasant. So to makes your life easier, we're introducing support in the emulator. So when you create an AVD, you can specify a virtual camera stream for the back camera, and you get the full 3D scene. And it works in any view that displays the camera stream. And using the keyboard and the mouse, you can just navigate around the UI. You even get patterns on that fake TV screen so you can do image recognition and pattern tracking, that kind of stuff. It's available today. Also we have Sceneform. So they already talked about it in the dev keynote. I won't talk about it too much. I will only say, this is what I've been working on for the past few months. So I still don't have a name, because we did that to Daydream. But anyway, that's what I was up to. And if you want to know more about Sceneform, there is a talk tomorrow afternoon at 5:30 PM. And we're going to talk about the underlying tech and how to use the API. And finally, Chrome OS on Pixelbook, you can now run Linux applications on your Chromebook. And in particular, you can run Android Studio. I believe there are some limitations. But one of the things you can do is run your Android apps as apps on Chrome OS. So you can use Chrome OS as a full blown Android development platform, I guess whenever it's available. [CLAPPING] Many, many more sessions today, tomorrow, and Thursday. You know, I said earlier, there's a lot of folks back home who have been doing all the hard work. We're just talking about it. So thank you very much to all the engineering teams, and the PMs, and the designers, and tech writers, everyone who made this possible. We didn't do anything. We're just here to hear you be happy. CHET HAASE: Some of us did even less. ROMAIN GUY: Yes. And that's it for today. So thank you very much. DAN SANDLER: See you at the Sandbox office hours. Bring your questions. CHET HAASE: Yes. We have Framework office hours. There is Android Sandbox running all week. DAN SANDLER: We've done our best to bring as many people from our teams here to be able to talk to you about these new features, these new APIs as possible, because we know you can't get that anywhere else besides Google I/O. CHET HAASE: And also there's an overflow area. I can't remember where it is exactly. Most of the speakers from Android sessions will make their way afterwards. So if you can't get your question answered at the session, you should be able to find them after that, like us there now. Thank you. [CLAPPING] [MUSIC PLAYING]
B1 中級 米 Androidの新機能(Google I/O '18 (What's new in Android (Google I/O '18)) 24 2 Tony Yu に公開 2021 年 01 月 14 日 シェア シェア 保存 報告 動画の中の単語