字幕表 動画を再生する 英語字幕をプリント SPEAKER 1: Remco was previously working at Boeing. And he got his degree from Brown University. And he will tell us about how to simplify dance models with the three buildings while still preserving legibility. Thanks, Emil. REMCO CHANG: Thank you guys for coming here. It's a great privilege to be at Google. So I'm here today to talk about kind of a broader research question that I've been working on. And the idea is try to understanding urban environments through the use of a thing called urban legibility. And I'll talk about that a little bit. So this talk is actually going to get broken down into about two parts. The first half is going to be on simplification of the urban models, while maintaining urban legibility. And this is a talk that I gave at SIGGRAPH this summer. So I apologize to people that were there. It will be a repeat, almost verbatim repeat. And the second half will be on discussion and future work, where I'll be talking about some of the things that we've been working on, as well as some of the things that we would like to be working on in the future. And before I start, I'd just like to talk a little bit about what we're doing at the visualization center at Charlotte. And specifically, one thing that we've been really interested in is in this idea of knowledge visualization. And to give you an example of what we mean by knowledge visualization, you can imagine like you have a screen of labels, there's lots of text, and they're kind of overlapping each other. Now, if somehow you can take this screen and extract some sort of higher knowledge, either from the scene or from the user, then it is theoretically possible that you can focus your resources on the things that you want the user to focus on. So in other words, we think of it as, you want to minimize the resources that you're using, but you want to maximize the information that you're giving to the user. And resource can really be anything. It can be CPU time, it could be a number of polygons. And at this particular case, it really is just a number of pixels that you're using on the screen. To give you an idea what we consider as a great knowledge visualization paper, here's something done by Agrawala and Stolte, who are both at Stanford. They're with Pat Hanrahan's group. And this paper is on rendering effective route maps. And here's an example of directions that would be given by MapQuest. And you see that this direction is physically accurate. It shows you exactly how to get from point a to point b. But that's about all it is. You don't really get a lot of information out of this. Whereas, typically speaking, if you ask somebody to give you directions, this will be something that people would draw you. So this hand sketch thing is not at all physically accurate. I mean, it's showing the highway 101 into a very small amount. Where if emphasis more on how you get onto highway 101, or 110, and how to get off. So of course, in their paper that was published at SIGGRAPH, 2001, they were able to mimic what humans typically do. And in this case, really showing off the important information in this task of giving people directions and maps. So we want to take this idea of knowledge visualization and apply it to urban models. So here's a model of a city in China. And the question is, what is the knowledge in this scene? What is it that we want to be able to preserve and to highlight? To answer that, we turn to this idea of urban legibility. And urban legibility is a term that was made famous by Kevin Lynch in his 1960 book called The Image of the City. So what he did for this book was that he went around the city of Boston, and he just asked local residents, and asked them to sketch out-- just kind of use a pen and sketch out their immediate surroundings. So what he actually got was a stack of these images that you see on the right here, where people just simply sketched out, you know, this is where I live, this is a big road around me, and so on. And he took this stack of sketched images, and he categorized the important things into five groups. He categorized into paths, which are highways, railroads, roads, canals. Edges, shore lines or boundaries. Districts, industrial, residential district. Nodes, which you can think of as areas where lots of activities converge. So as an example, Time Square in New York City. And then landmarks. And landmarks can really be anything. It can be a tree, it can be a post sign, it can be a big building. It's whatever people use to navigate themselves in an urban environment. So Kevin Lynch defined this idea of urban legibility as "the ease with which a city's parts may be recognized and can be organized into a coherent pattern." So that's kind of a mouthful. But to me, what that really says is, if you can somehow deconstruct a city into these urban legibility elements, we can still be able to organize a city in a coherent pattern. The use of urban legibility in computer science really goes back a little ways. Ruth Dalton, in her 2002 paper, just chronicles the history of what people have done in computer science in the use of urban legibility. And it kind of broke down to two groups. There's one group that tries to justify whether or not the idea of urban legibility actually makes sense. So what they did was, they tried to figure out if these elements are actually important to human navigation. And what they found out, interesting enough, is that paths, edges, and districts are very important to navigation, but landmarks are kind of questionable. There's some groups that think that it's very useful, there's some groups that say that it's totally useless. And the one element that's missing here is the element of nodes. And people have not really been able to successfully quantify what really a node is. So there hasn't been as much research done on trying to figure out if nodes are helpful at all. And the other group of researchers just use urban legibility, and in particular, in graphics and visualization. Most notably, Ingram and Benford has a whole series of papers where they try to use urban legibility in navigating abstract data spaces. So the question is, why did we decide to use urban legibility? And to give you an idea, here we take an original model. These are a bunch of buildings in our Atlanta data set, looked at from a top down view. This is what you would get if you use a traditional simplification method, such as QSlim. And I'm assuming people know what QSlim is. But what you see is that a lot of the buildings get decimated to a point where it doesn't really look like a building anymore. Whereas, our approach is a little bit different. We take an aggregated approach, and this is what you will get. And if we're apply a texture map onto our model, this is what you end up at the end. So, it's actually really interesting that when we take these four models, and we put in the fly-through scene, just kind of a test scenario, and we measure how many pixels are different from the original model. And this is the graph that we get. So you don't have to look at it carefully. But the important thing here that I'm picking out is that, basically, using all these models, they end up with very, very similar differences in terms of pixel errors. And what that says to us is that, even though you look at these four models and you say, well, they look very different to me. But in effect, if you measure it purely quantitative using pixel errors, they actually come out to be very similar. So what that really says to us is, we can't really just use pixel errors as the driving force behind simplification of urban models. We have to use something a little bit different. We have to use a higher level information in here. And to simplify this, let me just state, our goal for this project is to create simplified urban models that retain the image of the city from any view angles and distances. And as an example of what we get, you see the original model on the left. The middle image shows the model having been reduced to 45% of the polygons. And the last one is 18%. And you kind of see a little bit of a dimming effect across, when it goes from original to less polygons. But the important thing here to notice, that when you're doing this, the important features in the city are retained. So for example, the road here is still kept. The city square area is kept. And you pretty much still get the sense that this is the same city that you're looking at, even though there's only 18% of the polygons in the scene. I'm just going to run the application really quickly, and hopefully, nothing goes wrong. OK. So this is using the Chinese city data set. And this is running live. So as you can see, I can just kind of look around. Move to different places. And here-- hold on one second. This is where the demo goes wrong. OK. So I'm just going to start zooming out from this view. AUDIENCE: Can you mention how you got that geometry in the [UNINTELLIGIBLE]? Is that made out [UNINTELLIGIBLE PHRASE]. REMCO CHANG: Those textures are totally fake. AUDIENCE: [UNINTELLIGIBLE PHRASE]. REMCO CHANG: The geometry is actually real. So what we got was, we got the original footprint information, and we got approximate height information in terms of number of stories, or number of flights per building. And we estimated that each story is about three meters. So the geometry, it's kind of the extrusion of footprints. So it's not real in terms of the true 3D models, but the footprints and the positions are actually absolutely correct. AUDIENCE: Do you [UNINTELLIGIBLE] the fact that [UNINTELLIGIBLE] you get repeated texture patterns? REMCO CHANG: There is definitely some. But I'll talk about that in a little bit. Yes, sir? AUDIENCE: What kind of specification [UNINTELLIGIBLE PHRASE]? REMCO CHANG: As it turns out-- I'll get into that a little bit later, too-- this is-- AUDIENCE: [UNINTELLIGIBLE PHRASE] REMCO CHANG: Oh. OK, sorry. So the question was, what kind of hardware I'm running this on. And the honest truth is, I have no idea. But what I do know is that this is kind of the state of the art laptop from Dell. But as it turns out-- I'll explain this a little bit-- but the bottleneck's actually not in the graphics card. It's actually in my crappy code where I'm not transferring data fast enough. It's the pipeline that's actually the bottleneck right now. But that's just my fault. I wrote some crappy code. So here I'm just zooming out from that particular viewpoint. And to demonstrate my point, we're just going to keep zooming out and keep zooming out. Keep zooming out. And at this point, I'm going to freeze the level of detail simulation. I'm taking away the ground plane, I'm taking away the textures. And when I zoom back in, you see that this is actually what was rendered when you're that far away. And you can see the number of polygons at the bottom, right here. In terms of how many polygons was actually seen when you're really that far away. So the important thing here to note is that, even though we are doing some pretty drastic simplification, we still try to retain the important features in the city. And just to give another quick example of this, I'm just going to run this without the texture. We also take into consideration of high preservation. So what that means is-- aah. So I can be looking at a city from kind of an oblique view. And again, if I freeze this process, and I zoom out, you see that there's a lot more detail in the round where the viewpoint is. And as you go out further, the models are all greatly simplified. But the interesting here to notice, even for objects that area far away, the tall buildings are still rendered separately. So you kind of get that sky line, even when you're looking at it from different view angles, even close to ground plane. OK. Just to give you an idea about what work has been done in terms of urban flights or urban walk-through-- and people want to try different things. You know, visibility and occlusion is a popular one. And Peter Wonka has a great paper called Instant Visibility in 2000, and Schaufler has another great paper in 2000 as well, in SIGGRAPH. And in here, the idea is that if you use occlusion or visibility, you can reduce the number polygons that you're rendering, so that you can actually see a bigger scene, without actually seeing the bigger scene. And there are people who use impostors, or billboards with textures. Started with Marciel and Sirnley in '95. And then Sillion in '97 extended it to kind of blend between imposters, as well as real geometries. And in '98, Shalabi, his PhD thesis, extended Sillion's work, but added in some elements of legibility in there as well. They're procedurally generated buildings. I think this was really started by Peter Wonka in 2003's paper called Instant Architecture. In this year's SIGGRAPH, Pascal Mueller has a great paper on procedurally generated buildings. And then lastly, we just have popping, like Google Earth and Microsoft Live. So there are about five steps to our algorithm in order to preserve legibility. And in order, the first thing we do is, we try to identify and preserve the path and the edges. And we do that through this hierarchical clustering that I'll be talking about. The second step is in creating logical districts and nodes, and we do that through cluster merging. Then the third step is simplifying the model, while trying to preserve paths, edges, districts, and nodes. And that's done through our simplification process. Then we hierarchically apply the appropriate amount of texture, and that's through the texturing process. And these four steps combine to become the pre-processing step. And at the end of this pre-processing step, you end up with a hierarchy of meshes, as well as hierarchy of textures. Then we feed all this stuff into the run time process, where we try to highlight the landmarks, and choose the appropriate model to render, based on the viewpoint. And that's through the LOD with landmark preservation process. So first I'll talk about how we do preservation of paths and edges through hierarchical clustering. Here you see the result of two clustering methods. The one on the left is more of a traditional thing, like k-means and whatnot. And the right is what we use. And you can see the-- this is cool animation-- you can see where the two clusters meet in the first example, it doesn't really follow any sort of logical paths. Whereas, in our implementation, we really try to make sure that the two clusters are separated on a logical road. And the way we do that is by using something called single-link clustering. And that's a pretty simple idea. It's basically, iteratively groups the two closest clusters together, based on, in our situation, Euclidean distance. So as an example, let's say that you have six buildings to start with, a, b, c, d, e, and f. And you just start grouping them two at a time. And eventually, what you get back is a binary tree, or something called a dendrogram. The thing that's interesting here to note is that, in this particular scenario, the dendrogram is actually very unbalanced. On one side, you have node a. On the other side, you have b, c, d, and e, and f. And this doesn't work well at all for our purposes, so we had to do a little bit of balancing of the tree by penalizing larger clusters. And the idea here is, we want to create a more balanced tree. Here's some simple images of the first few steps of these hierarchical clustering processes. And you can see, red is one cluster, green's the other. And we just follow one particular path down this process. You can see that, throughout the simplification process, the divide between the two clusters mostly follow the road pretty well. So once we have the clustering, the next thing to do is to merge the clusters together to create logical districts and nodes. Here's an example of what we mean by creating a logical district. You have one cluster of buildings in red, the other cluster of buildings in yellow. And the blue outline shows what the merger will be. And this is what it should look like, where the blue outline should just encompass all of them together, including the black space in the middle. So the way we do that is pretty simple. We first find the footprints of each building, ordered in a counterclockwise manner. Then you find the two shortest edges that will combine these two things. Then you just start tracing it, starting with one of the vertices. Now, when we start with the magenta vertex, what we end up with is actually what we're looking for, which is the resulting merged hull. Now it's interesting to note that, when we start with a different vertex, and again, the counterclockwise fashion, what we get back is what we call the negative space. Or you can think of it as the error that's introduced by merging these two hulls together. And this is actually very important, because the negative space is what we end up using in determining what model to render later down the process. So once we have that, then the next step is to simplify it. And here's an example of what I mean by simplification. You see on the image on the left, it's the merged hull of the Atlanta data set. And here you have about 6,000 edges. And the idea here is, we want to simplify it without really killing too much of the features of the city. And in this case, we diminish it to about 1,000 edges or so. And I'm just going to show you real quick how this works. OK. So this is just a small data set that I'm showing. And right now, you see the blue outline, that's the merged hull of all of these buildings together. And as I start sliding the slider down, I'm actually doing simplification as I speak. So you can start to see that little features are starting to be filled in. And I just keep going. So what's interesting about this algorithm that we developed is that, eventually, if you just keep going at it, you get back to convex hull. So it's not at all the most efficient way of finding convex hull, but you do find the convex hull. Right. So once the polylines have been simplified, or once the merged hull has been simplified, we create what we call cluster meshes. these are nothing more than protrusion of the footprints, and where we determined the height of this cluster mesh to just be the median height of all the buildings in the cluster. And this is what we mean by what a cluster mesh is. Once we have that, then we apply texture onto all these cluster models. And we do it in the hierarchical fashion. So first of all, we give each cluster mesh six different textures. We give it a side texture. We give it a top down view of the roof texture. And we kind of do an impostor-ish kind of thing, where we give four roof textures from four different angles. And then we put these cluster meshes into bins, based on how visually important they are. And the idea here is that, if you have a particular cluster mesh that you know won't show up until it's like really far away, then what you can do is, you don't have to give it as much texture real estate. So, in such a case, what you would do is put it in the earlier bins. And I'll explain that in a little bit. Each of these bins that you see, n/2 and n/8, each bin will contain the same amount of texture resolution. So what that means is, if you have a particular cluster mesh that's put in the first bin, meaning the n/2 bin, the amount of texture resolution that it will receive will be a lot smaller, because there will be more people competing for it. Whereas, further down this pipeline, the more visually important clusters will actually get a lot more texture resolution. So this is a way for us to actually control the texture a little bit, because texture in a lot of these urban models, as you mentioned earlier, is a pretty big problem. So once we're done with all this, we put the resulting cluster meshes, and the hierarchy of meshes, and the hierarchy of textures, into our run time system, where we try to preserve landmarks, and choose the appropriate models to render. So first, we talk about how we choose the right model to render. And we start with the root node of our dendrogram. In this case, a, b, c, d, e, and f. And then we take the negative space of that particular cluster mesh, and that's shown here as an approximated 3D box, that's shown in red. And we project that box onto screen space. And if the number of pixels that occupies exceeds some user defined tolerance, then what we do is we reject that node, and we recursively check for it's two children. And it just keeps going until you find the appropriate cut in the LOD tree. Next thin I talk about is the landmark preservation process. Here you see the original sky on the left. The middle image shows without the landmark preservation. And the last image shows our method with the landmark preservation. And all you really need to do is you add a few buildings that's visually important, and you really give back the sense that the last image is very similar to the original. and the way that's done is, basically, projecting a user defined tolerance, alpha, onto each cluster mesh. And if there's any building in that cluster that's taller than alpha height, then it will be drawn on top of the cluster mesh. In other words, here's a scene of cluster meshes. And you see here, on the lower right hand corner, is a little green line. And that's the projected alpha height. And these buildings are drawn separately, because these buildings are taller than alpha height, so those will be drawn separately. And I talk a little bit about the results of what we get out of this system. Here you see this little line, blue line, that runs across the bottom of the scene. And that's rendering using the unsimplified meshes. And it's just a constant frame rate. And we see that, in this particular case, we have a fly-through scene, where the camera starts out really far away, zooms in to the model, and flies back out again. So what you end up seeing is, when the camera is far away on the top, you have really, really great frame rate. But once you start to really, really zoom in, to the point where you're almost touching the ground, then the problem is the overhead of traversing your LOD tree really catches up with you. In which case, the frame rate is actually worse than if you just rendered the entire scene independently. And correspondingly, in terms of the number of polygons that's rendered, you see the line on the top. And that's the number of polygons in the static, or the unsimplified, model. And you see that the two graphs are kind of are inverse of each other, because the number of polygons is inversely related to the frame rate. So in conclusion, there are a few things that we found by working on this project. The first one being that, when you just use per-pixel error, it is actually not all indicative of the visual quality of your simplified urban models. So what that means is, you really have to go towards some higher level knowledge, in our case, from city planning, to help extract and retain visually salient features in the model. In our particular case, using urban legibility, we find that it allows efficient and intuitive simplification of the models. There's some great limitation to what we're doing here. The first one is, this rendering engine that I implemented is horrendous. I cut all kinds of corners. So we're not using display lists, vertex arrays, any of the newer ideas with graphics card programming. So that needs to be worked on. The second thing is, the pre-processing step takes absolutely forever, because all the algorithms are basically n cubed processes. We can think of ways to simple it down a little bit, but it's something that we need to work on. And something that's inherent to this algorithm is that we end up with binary trees as our hierarchy tree. And binary trees says just inherently deeper than quad-trees or oct-trees. So traversing binary trees just takes longer. And the last thing is, we really want to improve upon this by allowing user interactions. And what we mean by that is, you can imagine that districts, in a logical sense, don't always follow roads or geometric distances. For example, if you have a historical residential area downtown, chances are, that's not bound by roads. I mean, that's probably just right immersed into everything else. But to the local residents, that district is maybe particularly important. So we need to have the kind of interaction where we allow the user to be able to manually put in going there, and say, no, this area is a little bit different. This is a historical region, for example. And that pretty much is it. So I want to thank the co-authors on this. Tom, Caroline, Zach, Bill from Charlotte, and my adviser from Brown, Nancy Pollard, who's at Carnegie Mellon at this point. And this project was funded by the Army MURI project. And some other people that I work with in the architecture department. So this is part one. And I'll just take some quick questions if you want. The second half, I'll be talking a little bit about what we want to do in the future and all. SPEAKER 1: So for the question, please remember that this talk will be available on Google video. So reserve all the questions with the confidential content for afterward. REMCO CHANG: Yes, sir? AUDIENCE: So you mentioned that the per-pixel error is not a good measure of quality. How do you measure your quality? How do you decide what's good quality rendering? REMCO CHANG: Do-do-do. You know what, we honestly have no idea. Sorry. The question was, if per-pixel is not a good indicator of the visual quality of the simplification, how do we measure our results? And the short answer to that is, we have no idea. I think it's interesting, similar to the paper by the Stanford group, where we did the efficient route maps, it's kind of the same thing. Because how do you justify that that kind of direction giving is good? Well, unless you really do a full blown user study, you won't really know. And I talked to some people about that particular project, and people agree that probably 99.0% of the people in the world would believe that that was a good sketch or drawing of directions. However, they mentioned some people living in the islands of the Pacific that really believe that the earth rotate around them. In which case, chances are that drawing would be absolutely useless to them. So it is very subjective. We thought long and hard about it. And I think, at the end of day, to be able to justify this as useful, you just have to do a user study. Yes, sir? AUDIENCE: [UNINTELLIGIBLE] how you merged the footprints of two neighboring buildings. That was in the two dimensional [UNINTELLIGIBLE], so how do you actually blend the geometry on two buildings in 3D? [UNINTELLIGIBLE PHRASE] REMCO CHANG: So the question is, the cluster merging process is all in 2D. How does that extend to 3D geometry? Short answer for that is, it doesn't really extend to 3D geometry. We basically assume that all the buildings are 2 and 1/2D. So all you have to do is play in the 2D realm, and do protrusion from that. And I'll talk a little bit about future work about the 3D geometries. Yes, sir? AUDIENCE: So [UNINTELLIGIBLE] working nicely, because you have [UNINTELLIGIBLE]. How would it work with like European style cities? REMCO CHANG: What's the difference? AUDIENCE: Well, they don't have [UNINTELLIGIBLE]. REMCO CHANG: Yeah. OK. Right. So the question is, how would this work for true 3D models, I guess? And again, the short answer is, I don't know. But I'll talk a little bit about what we plan on doing in the followup. Yes, sir? AUDIENCE: [UNINTELLIGIBLE PHRASE]. REMCO CHANG: Sorry. n is the number of buildings. Actually, let me take that back. In some cases, it could be the number of vertices. It's a terrible algorithm. n cubed is actually the absolute maximum, right? So in reality, it doesn't run at n cubed, but regardless, we never really tested the theoretical limit of where it really is running at. But that's something that we're definitely thinking long and hard about. Yes, sir? AUDIENCE: Yeah, just to put it in perspective, how many machine hours did it take to [UNINTELLIGIBLE] for demo? REMCO CHANG: For this demo, it was actually pretty quick. In this particular demo, there's about 30-some, 40,000 models. And that was probably, I would say, half an hour to 45 minutes. But I will honest with you, when we push it up about 50,000 to 60,000, I let it run over a week and it never quite finished. That n cube really catches up with you. OK. I'm going to quickly go over the second part of this, where I just want to pick your brains a little bit about what we plan on doing in the future. And so you guys can definitely help me out in terms of finding a good direction for this. Oh. This doesn't show up very well. This is roughly my research tree. And it's separated into three different categories. On the left hand side, we have mostly more core graphic kind of problems. The middle side is more on visualization of urban forms. And the right hand side is a blend between architecture, specifically urban morphology, and what we've been working on so far. So I start with talking about the architecture of urban morphology. The idea here is that, I just talked all this thing about urban legibility, and how we use of urban legibility for simplification. But the fact of the matter is, I never actually got back to saying well, what are the legibility elements in a city? We have no idea, where is the real road, where is a park, or anything like that. We just used the idea to do the simplification. Now, it would be really cool if we can somehow go back and find out, what are the elements. What are the important features to a city? So this we consider as kind of a feature extraction. And you can imagine that, if we overlay a model using very, very strong simplification, or a lot of simplification, we might end up with an image like this. When we overlay that on top of a map, now we can start to see that here's a road. And on the GIS data site, it might say that that road is Main Avenue, or whatever. And this area is something or other. Now we can start extracting, really, what are the elements that remain after our simplification? And we can do this-- oh, sorry, just pointing out some of the features here-- and we can do this in an interactive manner, where we start to do more and more simplification. And by doing that, more and more elements will start to appear. And eventually, you'll get all the way from the very, very important features, down to the very, very fine grain level of detail. And what that gets us is, now we can have a ranking, or a hierarchy, of urban legibility elements, starting with the most important to the least important. And this allows to be able to quantify and identify the urban model. In other words, this allows us to actually have some semantic understanding of the model that we're looking at. And this is important, because if we can start to do this, then we can start comparing between cities. So let's say, currently, if you ask any architects-- are there any architects here?-- OK. Good. If you ask architects, what is the difference between New York City, Washington, DC, and Charlotte? They would tell you that, in a very abstract and a very subjective way that, oh, New York City is made up of a grid. Washington, DC is more of a ray-like structure, where all the roads are emanating from the Congress and the White House. And then there's Charlotte that just has terrible urban planning. So the roads are just absolutely everywhere. But these are things that people can tell you in a very subjective way, right? But this is not quantifiable, this is not repeatable. And our goal is that, if we can somehow start to quantify a, city then we can start really comparing between these different cities in a more objective manner. And along the same line, we can start to track how a city changes over time. Again, in a quantifiable manner. Here I just show two images. Again, this is Charlotte in 1919. And then this is Charlotte today. And this is roughly the same region. But you can see that a lot of expansion has occurred. And nobody else can really tell you what, quantifiably speaking, what are the things that occurred over this hundred years. But what people can tell you is that, in 1919, Charlotte is a pretty perfect grid. Oh, didn't show up very well. But in 1919, Charlotte is a very good grid. Our downtown area is, you know, First Street, Second Street, Third Street, and so on. But once you start expanding, then the road just go all over the place. Now, we presented this idea to a workshop. And we found out that this whole idea of being able to track the city changing over time is still a fundamental challenge in the GIS community. People have not had a good grasp on how to be able to do this, in terms of a city changing over time. Along the same line, if we can start to quantify a city, the other thing that we'll be able to start doing is smart labeling. Or in other words, position-based intelligent labeling. So for example, if I give you a model, again, this is of China. Somebody says, I'm here. And then ask about, well, what are the important things around me? Now we can start to say, well, here's a canal in front of you, here's E Street, this is Main Ave, and so on. An interesting thing I want to point out is, now we can start to show the important features, based on their relationship to where you are. So E Street, for example, might not be very important in the global sense, but locally, to you, it might be very important. So we'll be able to highlight that because it's close to where you are. Similarly, Main Avenue, even though it's not related to you, because it's important, we'll show it. And then, you can start to show greater regions, such as downtown. And just group all the elements within it into a district. The second half of this is more of an academic question. Now, if we can actually do the thing that we just said, can we reverse this process back to it's beginning? In other words, can we start to sketch out a mental map similar to the ones that were drawn in Kevin Lynch's 1960 book? So we believe that there is a direct correlation between being able to do this point-based, intelligent labeling, and being able to retrieve a sketch of a mental map. So this will be a similar idea to what the Stanford group was doing with intelligent route maps. But instead, we'll be doing mental maps instead. The second half of this research direction is on visualizing urban forms. And this is a very interesting problem for the architects, because they mentioned that the study of urban form really hasn't changed at all since the early 19th century. People today are still looking at urban forms that are either 2D or 3D maps. So for example, this is something taken from ArcGIS, which is a product by ESRI. And this is what people normally look at when the urban planner is looking at an urban model. Now, if you look at this, and even though you overlay additional layers of GIS information-- in this case, you have some street information, you have some electricity information-- the bottom line is, these maps do not actively help the viewer understand changes or trends occurring in the city. In other words, it really doesn't help an urban planner to be able to do his job any better. So what we'd like to do is to start to apply some things that we do know as visualization people. We want to apply some of these visual analytics techniques to urban form. So let's say that we start to integrate a lot of these more abstract visualization elements, alongside this urban form idea. Now we can start to really see how a city changes, just in a more data kind of perspective. And by the way, this is not actually built. The background image, this whole system, was actually built for Bank of America on international wire transfer fraud detection. But it looks good together. And we showed this whole idea to the architects, and they immediately found things that they can start using. For example, the bottom image or the bottom window shows changes over time. and they can start to see how they can put in things that are changing the urban model into such format. So we believe that this is very promising. And the last part is more of a core graphics kind of question. And this I'm definitely doing a lot of head waving. There are a lot of people here who know a lot more about this than I do. But the first thing that we start realizing is, using our method, we can really do some model compression. And on top of that, to be able to progressively stream the changes from coarse to fine. This is really well built into the system that we have, so this should be a very simple step. But I think it's an important step, where not all the data has to be transferred all the time. You can just transfer incremental changes. And the second thing is, we want to be able to extend this to 3D models. Now this is a much bigger problem, because this includes true 3D models of buildings. And this also includes what we consider as trees and forests. Trees and forests is a very interesting topic in terms of GIS modeling, as such. Because people really haven't had a good grasp of this. Most people still do trees as billboards, or a combination thereof. And realistically speaking, people haven't really done a great job in terms of seeing a whole lot of trees. So to us, the problem is very similar. On one side, we showed that we can do simplification of lots of very simple 3D, 2 and 1/2D urban models. And trees and forests are the same way. People have been able to L-system trees and whatnot, where you can procedurally generate trees. But viewing trees in a large quantity is still a very difficult problem. So we think that we might give be able to use some of the ideas that we gained from doing 3D or 2 and 1/2D building models, and apply it to trees, and be able to take a stab at visualizing large quantity of trees, if not forests. So we took a quick stab at this. And what you're seeing here is basically a 3D model simplified using our method, and eventually get back to 3D convex hull. It's very preliminary. And the truth is, going from 2D to 3D, there's a lot of technical difficulties that we haven't complete figured out yet. But we're taking some stabs at it, see where it goes. And that pretty much is the talk. Is there any questions? Comments? Yes, sir? AUDIENCE: When you said that the landmark detection [UNINTELLIGIBLE PHRASE] you tried that with actual data sets [UNINTELLIGIBLE PHRASE] where most of the buildings are skyscrapers, or all different kinds? REMCO CHANG: Yeah. So the question is, in terms of landmark detection, have we done this in a scenario where all the buildings are skyscrapers? The answer is yes. And the idea there is, because the system doesn't really care about whether or not they're skyscrapers or not, the idea is, you basically have a base cluster mesh. Anything taller than that, based on your current view position, would be shown on top of it, or overlaying it. It doesn't really matter if they're all skyscrapers or not. They will still be rendered correctly. AUDIENCE: [UNINTELLIGIBLE PHRASE]. REMCO CHANG: That's basically based on where your eye point is. So that's a very good question, though. I can talk to you a little bit about this afterwards. But the idea of projecting your negative space, and the reason why the negative space is depicted as a 3D box, is because you do want to know about the height. So when you're looking at a tall cluster mesh from the side, the projection of the negative space will still project as something really large. In which case, it will be broken down into smaller sub-pieces. Anything else? Actually, just to follow up a little bit about that, there's some interesting question about what is considered as a landmark. And that's a very open question. If you ask a person who lives in the city, and you say, well, what is a landmark around where you live, they might pick out the local grocery store. They might pick out the watering hole, the pub that you usually go to. Maybe because it probably has been there for 40 years, everybody knows about. But from a purely graphics and visualization perspective, that information is not really as relevant, because they're not visually important. But if we're trying to do this system, for example, for the purpose of a mental map, then what constitutes as a landmark is not just purely geometry. Then you really have to start incorporating other GIS data sets to be able to figure some of those things out. Any questions? Any other questions? Well, thank you very much. I really appreciate it. [APPLAUSE]
B1 中級 都市の環境を理解するためには、都市の環境を理解するためには、... (Understanding Urban Environments Through the Use of...) 150 10 ya12081 に公開 2021 年 01 月 14 日 シェア シェア 保存 報告 動画の中の単語