Subtitles section Play video Print subtitles Cool. Hello, everyone. Hey. Good morning. Thanks for coming here. It's probably been a long week. The last couple of days have been crazy here, and the traffic's not been fun either. So hopefully it's been working all well for you guys. Some of you have made your way from a different parts of the country, many different areas. But I hope it's been worth it. And there's a lot of interesting things that we've been doing. So we want to talk a little bit about some of them. You've heard our keynotes and the great products we have, et cetera. Today I'm going to talk a little bit about machine learning, right? The title says machine learning is not the future, which is kind of weird given that we're talking about machine learning in everything we're doing here. But the point that I wanted to bring across here is it's not really the future. It's part of everything we have today as you're seeing without products as all the things that we're talking about. We're using machine learning today in everything we're doing. And I want to talk about some of that and tell you how you can do that too. But first, let's start with this all started. So and this goes back to 1997. I don't know how many of you remember this Tik-Tok from "The Wizard of Oz." He's one of the early versions of robots in modern times, really. Although there are references to automatons going back to the 5th Century B.C. And stuff, even the Egyptians and others way back when. But in the modern times, this is probably one of the earliest references to something that was close to a robot. Any of you familiar with Tik-Tok? How many of you are familiar with Tik-Tok? Not too many. Let me tell you a little bit about him. So this one is-- in fact, the term robot wasn't even coined back when Tik-Tok was created by Frank Baum. And what this robot would do is it had a winding thing. And you needed to wind the robot. And it would run. And it could do pretty much what a human could do but not really be alive. And it was a great reference to things that we've always wanted to do. And so a lot of this AI and robotics and a lot of these things have been part of science fiction for a very, very long time. In fact, a lot of what we do today in science has been driven by people, literally authors, et cetera, are able to really build up and think about what the future might be like. This is one example, but there are many, many more. For example, if some of you have read Asimov, he has a bunch of books. And he has this whole robotics cities, again way back in the '30s where he talks about robots and how they might be like and the kind of things that come with it. And that was very interesting. A lot of people over the years have been inspired by these science fiction books and movies to do a lot of interesting things. Another example of that is back in the '70s from "The Hitchhikers Guide to the Galaxy" there is this robot called Marvin, which is this depressed-- it's always depressing. He's always talking about these things he's not happy about. He's just too smart for all the things happening it. And it's really a great example of what people have been thinking about what's going to happen in the future. And now going to another example, "Star Trek," another thing. The example that I put here, you can go to Data in "Star Trek" who talks about-- this is a poem he composes, which I thought was pretty funny, and how people have been thinking about AIs and what they would do and what they would be doing in the future. In this case, again, it's like a robot, like a humanoid, somebody who can do a lot of things, but is not quite human. Then coming to more recent times, more closer to 2001, if you've seen Steven Spielberg's movie "AI" called "AI" itself, they have these mechanical robots that can really do-- that look like humans, that can do all kinds of things. But they don't feel. And then they build this little kid David, who can actually feel and love as well, and really changes how the perception is and how they think about AI or what it means to the people around them. And then much more recently there's this movie called "Her" in 2013, which talks about AI without the shape, without the robotics, and all of that. But AI is just that live within your computer as an OS, essentially, that can interact with you. That's your assistant, but much more than that. It actually feels as well. It understands things, et cetera. So all these are great. There are so many different things that people have been talking about, people have been thinking about. But these are still science fiction. This is not really what machine learning is about today. This has always been the future. It's still the future. Maybe at some point it will be a reality. But that's not where we are today. That's not what we're talking about. But there are some real things that we can do today. And those are some of the things that I'm going to talk about today. So we've actually made over the last few years real progress in terms of all the different things that we can do with AI with machine learning, use all the products that you see around you. And I'm going to talk a little bit about those. There's so much in there that has benefited from machine learning from what we call AI as well. And so this is some of the smallest of products, really, at Google that we made that used machine learning in some ways. But at Google, anytime we think of a product there's, of course, programming. And you build it, and you do all sorts of things with it. But machine learning is an integral part of everything we do in building that, because we want these products to be really smart to give you the right things, to not just follow your actions, but really give you the right things when you want them as you want them. And I'll go over some examples of these in later slides as well. So before I go into other things, let's just go a little bit into what deep learning and machine learning is about. And I'm just quickly going to give you some examples on a website that we have. So part of this slide is going to talk about TensorFlow. And recently we put the site up called playground.tensorflow.org that allows you to play with neural nets that allows you to really do different kinds of things, and allows you to really understand how these networks work, how machine learning works, and be able to play with some of those problems. So I'm going to start with a very, very basic problem classification. The goal in this case is there are two kinds of clients. I just want to classify that it's ARB, in this case the orange or the blue ones. And I'm going to use the very simplest case. It's a very simple linear classification, if any of you guys know what that is. But the idea is you have some inputs. In this case, the x-axis and the y-axis are inputs. And based on those two inputs you want to decide if it's a blue one or an orange one. And so what this model is going to learn is some parameters to figure out-- OK. If I get some point, I get the an x1 value and a x2 value, how do I decide that? It's some mathematical computation that it needs to do. So in this case, it does that iteratively. Let's just quickly run this. I've just set it up as a linear model. So lets just run it very quick. And as you see here, its looking at x1 and x2. And very, very quickly it figures out how to separate those out. It just draw a line in the middle. And its just optimizing what that line looks like, how to separate these out very simply. So this was a very easy problem. This is typically not how the real problems are, but it's a great place to start. So let's go to a slightly more complex problem now. So let's go to this one here where you still have those two kinds of points, but now they're clustered differently. So there's one in the center-- is a circle, and the rest are around it. So can we actually solve this using the same kind of methods? Can we actually classify using the same thing? If I run this, do you think it would work? Any guesses? Let's try it out. So it's trying to separate this out. It's basically trying to draw a line. But it's lost. It's really not making a progress if you see what's happening here. So there are a few things you can do about this. One, in terms of classification, again, going a little bit into machine learning into the details, you can add more features, more kinds of things about the input data that helped the model understand or how to separate them. So if you guys remember your math back from high school or college, in this particular case, given how these clusters are structured, if you looked at a couple more features from based on the inputs, like in this case, x1 squared and x2 squared, you can probably use those to separate this out better. So let's try this. Let's add these two and see what happens. And taking those two values, now it knows how to separate those and clearly separate it out into two separate, different classes. So this works. This great great, right? This is what happens or used to happen in machine learning very often for every single thing that you do. If you take a bunch of inputs, you want to solve some problem. It's typically the inputs themselves that are not in the same kind of mode that you need to solve the problem. And what you do is called feature generation where you are combining or crossing those features in different ways to really solve your problem. But there's a lot of work you need to do in figuring out the right features that make sense. In this case, it was a simple problem. So we knew x squared, x1 squared or x2 squared would work. But in some cases, the problem is not as simple and you really need to do a little bit more. So for those kind of cases, we have something called deep learning which seems to work really well. So let's go to go back to this example and let's see if we can solve it in a different way as well. So let's take out these two features first. Let's change this linear to a nonlinearity. In this case, I'm going to pick correctify linear unit. And I'm going to add a hidden layer with a few more neurons. So what this is doing is it has some inputs, as we said in this case just x1 x2. I'm adding another layer, which is basically trying to combine those inputs in whatever way is interesting to solve our particular problem. And then let's see if we can train this to actually solve the same problem as well. So if you see this, we took the same features, no feature generation. We just added one layer. And this is a simple problem, so one layer works. And it's very easily able to separate those out. This is essentially what deep learning is doing for you. As you add more layers, it can understand and learn the right layers of abstraction that are the most interesting to solve your current problem. And of course, it depends on the problem. It doesn't have to be a simple classification. It could be regression. It could be recommendation, whatever you are trying to do, really. You can apply similar techniques to do that. So let's go back to our slides. In a sense, deep learning's basically the same machine learning ideas with more layers there that allow you to get different kinds of abstraction. Of course, as part of that you get many, many more parameters. You learn things differently. This example is actually for what we call inception retreat, which is our state of the art image model for image classification. So And we'll go through some examples of how we use it. But this basically gives you something nearer human accuracy. So machine learning has been around for a long time, in fact way more than 10 years. What's really changed since then? One of them is more compute, definitely. Over the years, our computers have gotten faster. We have more of them. We've been using newer chips, CPUs, GPUs, now 2PUs here. And so a lot of things that we're doing here-- and more computer allows us to really build more complex models, more bigger models. But that needs a lot of other things to go with it as well for that to work, in this case, better algorithms and more data help as well. So over the years we've improved the algorithms slowly. A lot of these algorithms that we talk about in deep learning are 30, 40, 50 years old. But over the last 10 years, we've seen some improvements in the right places. For example, in a case of deep learning, there are some improvements to back propagation or the kind of nonlinearities that we use that really make it much easier to optimize those. It used to be really, really hard to train some of these models, even if you had the compute. Even if you had the data. Now it's much easier and there are better techniques to do that. So the math and the algorithms is improved. And of course, the amount of data that we have, that keeps growing. And that really is something that machine learning can help. If you think about it with small amounts of data humans are great. You can maybe read 10 pages of a book and understand that and sift it down. But what if you had a million pages? You can't really expect a human to go through those million pages, understand those, and summarize it for you. You really need some machines that can automatically do this for you. And that's really where machine learning comes in. So now let's talk a little bit about what's happening in research in machine learning now. We've talked a little bit about science-fiction, mentioned a bunch of products that are using this. And I'll go a little bit more into the products later. But let's talk a little bit about the research, where this is going, some of the examples of where things of made improvements in the recent past. So the first one I want to talk about is AlphaGo. How many of you of heard of AlphaGo? Wow, quite a few. So if you remember actually almost 20 years ago now, there was deep blue, which is a similar chess computer. In some sense it was similar that the goal was, OK. Can we really beat the best humans, Casper of the reigning champion in that time using computers? And back in 1997 was the first time Deep Blue, which is this computer built by IBM that actually beat the world champion, beat Kasparov in a full championship match. That was pretty amazing. And that was 20 years ago. It took a long time to go from there to Go, partly because Go itself as a game is way more complex than chess is. The possibilities that you have at every level at every step are much, much bigger, much more than what chess is. And so one of the differences there is with Deep Blue you could get away with some smarts and a lot of brute force. You could have a large supercomputer. You have some basic smarts and a lot of rules around-- there are a lot of chess players there that actually taught Deep Blue in the sense it would train them on-- OK. Here the good moves. Here are not, et cetera, et cetera, plus allowing it to do a lot of brute force in trying out-- OK, looking ahead how many different moves you can do. Now those same techniques just could not work for Go because of the different possibilities at every step. At every step we can do a really large number of moves, and supporting that or being able to go past a few steps is extremely hard, even for modern day computers. So what really worked in this case was a combination of smartness and the computer. And in case of smartness, they applied deep learning to figure out, OK, looking at a board, what kind of moves make the most sense? And then let's try them out and see what makes sense and which ones of these play well. And so it's a combination of improved algorithms, improved machine learning, combined with the compute power as well. Another example is ImageNet, which is this data set of a million images with 1,000 classes of different kinds. In fact, a lot of classes are so close together that it's very hard even for humans who don't know about them to identify without learning. For example, there are breeds of dogs which look very similar, at least to me. For people who really know dogs, of course, it's easier. Over the last few years, over the last five years, especially on this data set, this as really driven the state of, well, vision understanding for machines is. So about five years ago, the accuracy that computers would get was maybe 70%, 75%. This was the measured top five accuracy, which is basically is the most-- the correct answer among the top five results that the computer predicts? Now, five years ago this was something like 75%. Today it's past. It's over 96%, which is considered to be better than human accuracy. Humans also make mistakes in this kind of thing. So it's really coming a long way and just shows the kind of progress we can make if we really get down to improving machine learning. This one's an interesting one, taking the kind of image models that we've trained and really letting them dream. This is what we call Deep Dream. There are some folks at Google who basically took those new networks and wanted to understand what's happening in these networks. What are they learning right? And so in this case, what they did was took a few neurons in the network, just let them train from that point and try to see what we can generate from those neurons, what they might be understanding from the images. And so they converged to some of these. So in some of these images in the top right, it's basically started from some kind of palace or whatever and combined those, and [INAUDIBLE] of that. The one in the top left it has some fountains in there and a bunch of other things that sort of combined. And this is very interesting because in some ways there are similarities to what people do as well. Of course, this is a very different view. But it's starting to learn in many, many ways. Machine learning is really doing the same kind of learning that people might do. So now I'm going to give you some examples. Let me just go to this and play some, a couple more videos for you. In this case, it's doing what-- it's basically taking the video and applying different kinds of painting styles. In this particular one it's just picking a blue painting style and applying it to the existing video in real time. Now it's a charcoal sketch. This is just an example of how we're mixing machine learning with things like art to all kinds of interesting things in and interesting combinations things that you can do with these. Here's another one where it makes it look like a newspaper. So this one's based on a paper based on by Justin Johnson and others here at Sanford. Let me give you another example of similar things that we can do. So this one is interesting where the network learned from a bunch of Chinese characters. It looked at how they're drawn, and was able to really draw completely new Chinese characters I've never seen before. These aren't real characters, actually, for those of you who understand the script. I don't, but then they look very real. Even for people who understand the script or who know the script, they feel like they're almost there, they're clearly not the right character or anything that they've ever seen. But it's very understanding what's happening there. It gets tagged it's supposed to look like this and is able to understand that whole domain as well. I'm going to show you a couple more videos. Here's one where it's basically learned from a lot of numbers. These are house numbers from Street View. Let's just go back here and play it and pause it at the end. And so it's learned a bunch of-- it's looked at house numbers from Street View and is able to really generate those. It's actually never seen these particular numbers. The network is generating them. This is the kind of things that these networks can do today. And the last one, in this case, it's actually generating faces. And so, again, it learned from a large face data set and now is able to generate faces in all kinds of ways. Again, these are not faces that's seen directly. It's regenerating them based on all it's understanding of what faces look like. Yeah. Another one that we've actually released very recently is this model to parse sentences. So this is a very semantic parsing of sentences where it looks at a sentence. In this case, I booked a ticket to Google, and tries to understand what the subjects are, all the grammatical stuff about that, actually things that I personally don't understand, and it's hard for me to learn. But this is able to actually learn from a lot of sentences that were given to it. In this example, it's actually making the sentence piece by piece, and then building a tree of hierarchy, and doing the parsing in this case. This kind of thing is extremely useful when you're trying to understand natural language. And this just goes to show how much these computers can understand the language that you and I speak in some cases better than humans can as well or at least better than kids can. Here's another example, sort of going back to the data example where in science fiction this-- here's this sort of poem that a computer might generate. This is actually something that was generated by a real program today or a real model today. It's clearly not great, but in this particular case, it just learned from a bunch of books. It read those books or understood those books. And was able to generate this particular poem. So now let's talk about this grateful research. Now let's talk about a few products and how we're using them. So we had this slide early on where I showed you a bunch of products that [INAUDIBLE] using this. Let's go into some of those products. Let's see how we are actually using them. What are the kind of things we're doing with machine learning in these products? First a search-- I'm sure most of you have-- or probably all of you have used search or are using search in some ways. This is a very good example of how machine learning is used. The slide that I showed you here, there are at least two ways we are using machine learning here. The first, if you use Y Searcher-- so if you click that icon on the right for the mic and you say something, there's machine learning being used to understand what you're saying and converting it to text. So you say something. That's really a bunch of bytes for the computer. To convert that into actual English or whatever, like you're speaking, it needs to understand that. And there are pretty sophisticated machine learning models that do that for you today. Then the next part is once you have that, what results to show you? Google has billions or trillions of pages really indexed, which you probably don't even think about. What you see typically when you do that search is just 10 pages on that first page. And you really want those 10 pages to be the best ones, the most appropriate ones for you. In some cases, you might go to the next page or so. But most people are just going to look at that first page and make their decision. So how do we really take those billions of pages, match them to your query, and then even then it's probably millions or maybe more, and then put them down to those, whittle them down to those top 10 pages? And that's, again, machine learning at work in many, many forms. We're taking all these pages, understanding your query, matching it through all the data that we have. Maybe if you're looking for a restaurant near Mountain View, we're going to leverage that and understand that that needs to be combined with the kind of pages that are there. If it's something else, again, understanding the query and the documents together, combining them, and then ranking them, and sorting them, that something that machine learning does really, really well. This example is of pictures. We talked about vision models. They're actually used in everything you do now. So for example, over the years, image collections, the amount of photos I have are huge. They're just growing. With digital cameras coming on, it's been so much easier to take pictures and just keep collecting them. You never throw them away. But then when you want to find a picture and you really want to go through those collections, it's really, really hard. So how can you do that? In this case, the same kind of research that I was talking about, which understands those images, is really indexing and labeling all the pictures you have in Google Photos. And in this particular example, if you search for cherry blossom, it really shows you all your art pictures of cherry blossom, and can display them for you. Another example is email. So there is this thing called Smart Reply, which we launched a few months ago, which you might be using in email. And what happens there is you get an email. You typically, in this case, you probably just want to get a quick, easy answer. And it just gives you some choices. Here, pick this one. And what it's doing is it's really understanding what your email thread is about and then suggesting a bunch of replies ranking them whatever makes the most sense and showing them the most interesting ones to you if it makes sense. Another one is Google Music. There are many things happening in there. And one of the things that we do is we recommend a lot of things that you might want to play, maybe albums that you might want to play, maybe actual songs that you might want to play based on the time, based on what you are trying to do right now, based on your interest. It's really understanding all of that, sorting them, and, again, ranking them, recommending the right things for you. We only want to make it easy for you to find what you want. The ideal case would be you go there. Let's say it's afternoon or let's say you're driving back from work. And you typically like to listen to, say, classical songs at that time. It should really just recommend you the classical song and maybe the top station that you do every time. You shouldn't have to look for it just because-- just before this, you were doing something else. So all of these are examples. Machine learning can really make a difference and are helping us in our everyday lives. And one more that I'm going to talk about is spam. So everybody uses email. A few years ago, many 5, 10 years ago, spam was this huge issue for everyone where most of the emailing in our inbox is starting to be spam. You have to actually filter them out manually. So it's not that from that time on to now spam is reduced. In fact, spam continues to increase. It's just there are filters, and algorithms have gotten a lot better at recognizing what's spam. Again, they're not perfect. Occasionally you'll still see spam. Or maybe we would say something is spam when it's not. But they're getting so much better, that I don't go to my spam for spam box and see what I might have missed, or the other way around. I rarely get emails that's actually spam, even though if you actually read the numbers about how much spam emails there are, these are still way more, maybe 10 times more than the real emails of spam itself. So lots of improvements in machine learning, lots of improvements in the products that are driven by machine learning. So it's not about the algorithms or the technology itself. It's about how do we apply it to the products that we build? How do we make things better for all the things that we're trying to do? All of this technology, and algorithms in the computer that we're building, it's all towards solving real problems, solving real needs for things that we do today. So now let's talk a little bit about TensorFlow. That's the product I need. And it was built to help push machine learning forward. Our team is a research team. And part of the goal for us was to really push machine learning forward. And that's why we built it. But that's not the angle, right? We want to take that machine learning as well and apply it like I was saying in all of these products. TensorFlow allows us to do that, and now you as well. And I'm going to tell you of how we can do this. So it is open source. It's a machine learning lab available on GitHub. It's, in fact, the most popular machine learning library we opened source six months ago. And it's rapidly going to the top. There's a lot of interest. There are a lot of people using it in all kinds of areas. Some of the research that I talked about, actually people outside Google who've taken up TensorFlow, and used it in doing a lot of those-- building those new things and new ideas, people are including it in different kinds of products as well. So at the core, it's a library that's-- let's go a step forward. So what's this library for right there? There are many different things that we can do. Like I was saying, our group is doing a lot of research. So it is for researchers who want to take machine learning and really push it forward, make it much better than it is, improve the algorithms, improve how we can do things with it. So it is for them. But then there are data scientists who want to really take those algorithms-- let's see our data, let's see our product. And you want to improve that product by playing this machine learning to the data that you have. And so the data scientists can take those algorithms because they're also open now. They're made available to you on all kinds of places. You can take those. You can use those and make some improvements. And then there are developers, like a lot of us here, actually, who are trying to make, say, an app, and want to make it better, or say, apply a product to their data center. And TensorFlow allows you to really take all the work that researchers and data scientists have done, and really run those same models, the train models in production using the exact same APIs without having to worry about-- OK. This seems great, but how am I going to deploy this to production. That's another complexity that that's harder. So we definitely try to make sure that from the beginning, from the point that you get started in machine learning, to actually pushing it out, to production, it's a very short time cycle. You shouldn't be stuck because of infrastructure. That's taken care of you. That's taken care of for you. And you can focus on what's important for you in the most ways, which is really building the products, helping your customers, helping your users. So just a couple examples of TensorFlow-- not going to go too detailed into the code, but this is a very simple example where we just take two inputs, do a matrix multiply, and just print out the result. So what it's doing is in the first two cases, it's creating two nodes in the graph. And actually, let me just-- yeah. It's creating two nodes in the graph, and really doing a multiplication, which is another node, which takes the inputs from the other two nodes. And then the last one really evaluates based on the two inputs, combines them, and generates the output. This one's actually a regression problem. So we talked about classification earlier. In this case, it's trying to predict some values based on, can we have some inputs? Can we have some outputs? You want to learn to predict the outputs. It's basically doing something similar to what we were seeing on the playground where you're taking inputs, you're multiplying by matrix, and you haven't lost it. You're trying to optimize it, which is a typical machine learning thing. And then you just loop through. So the last, at the end, you basically have these. You're going through a bunch of steps, and say 200 times in this particular case, and you just print this out. And over time you're basically learning to predict things better. Again, very simplistic examples, typical models might be much bigger. But in a lot of cases, you might also be able to use models that are already there. Why don't we just give you some examples of the kind of performance we provide comparing it to some standard benchmarks that are there on an open source website. And early on when we launched TensorFlow, it was in the fastest library out there. And there were a lot of improvements that we had to do. But over the last six months, we've really focused on that. And in this point, it's really at the top in terms of compare compatible to everything else out there. And we continue to improve it. There's not the end of it. There are many more improvements that you can do. In this case, we're combating four different kinds of models. In this case, they happen to be all image models. And as you see, the load is actually better. In a few cases, it's just slightly worse. In few cases it's slight-- in one case it's likely better, and so on. So it's roughly about the same ballpark performance at this point. Now, one thing you can do with TensorFlow as well is you can run it across many, many machines rather then running on just one machine. So the previous benchmark that I was showing you was basically running on a single GPU and training these models on GPUs, something that a lot of people do. In this particular case, we're training on many, many more machines, so each machine having a single GPU going all the way to 100. And you can see it scales quite well. In fact, up to 16 it's almost linear. And then it's still going to be used to scale you in past that. And this is something that we use internally as well to train really large models and large data sets, which is-- as the data is going, it's really important to be able to do this, to be able to run on clusters, which many of you probably do already. Oops. So next I'm going to talk a little bit about TensorFlow for Poets. So there's actually a code lot for this as well if you want to look at it and look at it online, or I believe there's a place here too. So the idea here is-- let's say you have a bunch of images, in this case flowers, and you want to learn to recognize those flowers, or maybe you just identify if it's a flower or if these are five different kinds of flowers. And what you can do is you don't have to have a million examples. You might just have 1,000 examples, lets say. But you can get started from a model that works really, really well. For example, our best image model that you've open sourced and made available, and that's what this example talks about. Take the top part of it, just train that on your data set from the point that it's already trained on. So it's just going to learn to classify the flowers. It already understands image well, so it doesn't have to really know what kind of features are important for classifying images. It knows that already. It can just take your particular models, your particular data, and learn to just separate those out into different classes. So there are a few tools that I would like to mention at the end that you can play with, or that you can use before you go in terms of just making things work well for you. So a lot of you are probably familiar with Apache Spark, which is used for doing analysis and doing learning on large data sets. So there is an integration of Spark with TensorFlow-- somebody at Databricks did that-- where you can actually take some images. You can create. It's very simplistic in the sense you can create a lot of different models and really use it to pick the best model. You do that a lot because there are some having parameters like, say, learning rate or how big should your model be, et cetera. You maybe try 10 different models and use that to pick the best one. Another thing I would like to point you do is Google's own Cloud Machine Learning Platform. We announced this a couple of months ago. It's under limited preview right now, but you can sign up, and we will make it accessible to you soon, where you can really have the same power of TensorFlow that you have on your machine say, but maybe don't have the compute resources, or you don't want to manage all the resources that you need to really run at large scale. And cloud takes care of all of that for you. So you can experiment on your machine, make sure you understand what you want to do, and then push it off to the cloud, let it take care of running large scale stuff. There are a few examples of what you can do with models that have already been trained by others and really take them and just apply them. There are some APIs on the forum today for what's called the Mobile Mission APIs, which allows you to do face tracking. So let's say your pictures, it allows you to-- it identifies parts of the faces. It first identifies the face, and then parts of the face like the eyes, nose, et cetera, so you can do different kinds of things with it. It also recognizes barcodes and QR codes and automatically understands and parses them for you. On the cloud there are also some APIs for vision and speech in translation, which you can basically just send your image, and it tells you what kind of image it is among a lot of different classes. For speech, again, you can upload data, upload speech, and it will give you back text for whatever kind. For translation, you can give it text in one language, say English or Japanese or whatever makes sense for you, and it's going to convert that to another language. So many different APIs, and this is a growing area that you are going to need to watch out for. So now for the future. It is really. I started by saying it's really not the future, but it is. It i the present and the future. I just wanted to call out that it's important for you to start thinking about it now and not wait for the future, really. And it's really your turn to start thinking about how you can incorporate machine learning into everything that you do. What people are looking for today is not just applications that do what they want. There are tons of things that are already doing that. How can you add that edge? How can you how you make it really better for that user? By understanding what they want before they ask for every single thing. And that's really something that you can do. Think about this as you build your next application, as you build your next project. And I think that's how we'll change the future. That's how we'll make machine learning part of the future. So I wanted to leave you with some place with some pointers to where you can read more about TensorFlow. It's on GitHub, and their docs are tensorflow.org. And going back to my reference for Tik-Tok, Tik-Tok is waiting for you to really make the difference to do these things now. And thank you. [APPLAUSE] [MUSIC PLAYING]
A2 machine learning learning machine spam case data Machine learning is not the future - Google I/O 2016 115 10 Amy.Lin posted on 2017/04/13 More Share Save Report Video vocabulary