Placeholder Image

Subtitles section Play video

  • What's up, Josh here.

  • So today Google wrapped up IO 2024, where they announced several new AI powered features and integrations and generative AI that are really, really exciting and there's a lot going on.

  • So in this video,

  • I'm gonna attempt to break down the two hour long presentation into something more manageable and hopefully help you figure out what's going on and how you can take advantage of these features for yourself.

  • So yeah, go ahead and drop a like, subscribe and let's get into it.

  • Now, if I could summarize this presentation into two sort of main buckets of features, it would be number one, integrations and number two, long context.

  • Now, integration is something that really only the Googles and Apples of the world can do.

  • I'm talking about companies that own different products that seamlessly integrate with each other, which makes it easy for them to sort of integrate a new feature, such as AI into all of these products.

  • So for a big portion of the presentation,

  • Google really showed off a bunch of ways that they're now integrating Gemini into all of these products, which in all of the demos is basically some variation of helping you find and organize information.

  • Now, one thing I'm sure we all deal with is emails.

  • And one of the most impressive integrations with Gmail was being able to go through your entire inbox to organize and track, let's say, receipts.

  • So in this demo, it found 37 receipts and also created an entire spreadsheet that really would have taken hours and hours to create if you were to do it on your own.

  • Then once that spreadsheet was created, we also saw Gemini analyze that data, even creating a graph to visualize everything, which is pretty neat.

  • There was also another example where let's say you missed an entire email thread.

  • Well, now you can just ask Gemini to summarize the entire email thread for you.

  • And then based off of that summary can also help you draft up an email.

  • It can also summarize an entire recording of a video conference up to an hour long, as long as it's on Google Meet.

  • Another example is within Google Photos where they rolled out Ask Photos, which allows you to just search for photos just like you would on Google, except it's your own library.

  • Say you're at a parking station ready to pay, but you can't recall your license plate number.

  • Before, you could search photos for keywords and then scroll through years worth of photos looking for the right one.

  • Now you can simply ask photos.

  • It knows the cars that appear often, it triangulates which one is yours, and just tells you the license plate number.

  • And then within the Google Workspaces suite, they're rolling out side panels, which again is just another bridge to Google Gemini, which is just this floating window where you'll always have access to Gemini to search through your documents or even summarize them for you.

  • So like I said, these are all examples of integrations within Google's existing products that allow you to organize and find information faster.

  • But then something a little bit interesting here is that they're integrating Gemini into Google Search.

  • And so you can actually enable this feature right now, and I think the presenter said it best about what this feature actually does.

  • And Google will do the Googling for you.

  • So yeah, she said Google will do the Googling for you, which is just ridiculous to say.

  • But yeah, the new Google Search powered by Gemini now includes AI overviews, where it gives high-level summary of the results with suggested links for you to check out, as well as multi-step reasoning where you can ask it really long and specific things, like for example, a yoga or Pilates studio in your city that's highly rated and within a half hour walk.

  • And then it'll give you a list of studios for you to potentially check out.

  • You can also ask it to do a bunch of things like meal prep, plan a trip, or create a workout routine.

  • And you might be picking up on the fact that this is sort of blurring the lines between Google Search and Gemini, because they sort of have overlapping functionalities.

  • Now, I think the main difference between Search and Gemini is with Search, you're still getting human-generated like articles and blog posts and that sort of thing.

  • So for example, if you look for a meal plan on Google Search, it's still gonna give you human-created recipes.

  • Whereas on Gemini, it's just gonna give you meal plans that are written up by AI.

  • And overall, I think Search is still an extremely important product for Google to hold onto and sort of prolong the life of, because it is how they generate most of their income.

  • So yeah, that is the first majority of sort of like the bulk of what they announced is all of these different integrations into different apps.

  • But then the other thing that Google really leaned into is long context.

  • So in the presentation, we heard this word tokens a lot.

  • And while I still don't really understand what a token is, it seems to indicate that it is a chunk of information.

  • And the big thing that Google kept repeating is their support for up to 1 million tokens in Gemini Pro.

  • This basically means that Google's latest model is gonna have way more support for storing information, which is gonna be extremely useful for things like research, handling long documents, lines of code, and even analyzing video.

  • And yeah, this is really how Google is able to power a lot of the innovation that they're talking about, which includes looking through tons and tons of emails and documents that might be in your Google Drive.

  • And they also announced a couple of experimental apps that you can actually play with right now.

  • So there's this one called Notebook LM, where users can upload tons and tons of documents, charts, diagrams, and then have Gemini generate things like study guides, FAQs, quizzes, and even an AI generated like podcast type thing to help you understand concepts a little bit better.

  • Let's dive into physics.

  • What's on deck for today?

  • Well, we're starting with the basics, force in motion.

  • And that of course means we have to talk about

  • Sir Isaac Newton and his three laws of motion.

  • Ah, yes, the foundation for understanding how objects move and interact.

  • Join.

  • Hold on, we have a question.

  • What's up, Josh?

  • Yeah, can you give my son Jimmy a basketball example?

  • Hey, Jimmy, that's a fantastic idea.

  • Basketball is actually a great way to visualize force and motion.

  • Let's break it down.

  • Another app is called AI Studio, where again, you can upload entire research papers, code repositories, videos, photos.

  • And the way that I almost look at it is you're creating your own database for your own Google search, in a sense.

  • I think it's gonna be especially useful for researchers, students, and analysts.

  • These people who deal with tons and tons of data and documents and spreadsheets, which would otherwise take hours and hours to scrub through can now be done in literally seconds.

  • Now, I actually tested this for myself on a very small scale.

  • So I basically transcribed the entire Google I-O keynote and uploaded it to Gemini and was able to ask it questions about the presentation, which was great for research.

  • Which by the way, that entire transcript, according to ChatGPT, was only 126 tokens.

  • So it really just goes to show how crazy one million tokens is.

  • It really isn't that necessary of a feature for the everyday user.

  • Now, I think the really weird thing with Gemini is that if you're on the free plan, you actually cannot upload documents, even though with all of these other experimental apps that are free, like Notebook LM and AI Studio, you can.

  • And considering that OpenAI's ChatGPT 4.0, which is now made completely free, does allow you to upload documents on top of photos,

  • I really think that Google should add that functionality to the free Gemini plan.

  • Now, over on the mobile side of things,

  • Google announced Project Astra and gave us this really early look into this live interaction with Vision, which sort of reminds me of OpenAI's demo with 4.0.

  • But yeah, in this prerecorded video, the presenter sort of walked around the room, pointing the camera at things and asking questions along the way with real-time responses.

  • What does that part of the code do?

  • This code defines encryption and decryption functions.

  • What neighborhood do you think I'm in?

  • This appears to be the King's Cross area of London.

  • Now, all of that is very impressive.

  • However, pointing your phone around is not the most ideal, so they showed this prototype off, which I thought was pretty interesting.

  • Do you remember where you saw my glasses?

  • Yes, I do.

  • Your glasses were on the desk near a red apple.

  • What can I add here to make this system faster?

  • Adding a cache between the server and database could improve speed.

  • So yeah, there may be a Google Glass resurrection maybe a couple of years down the line.

  • Now, in a really confusing way,

  • Google also teased Gemini Live, which seems to be the actual thing that consumers will be getting in a couple of months.

  • So this is, again, a live conversational feature built right into Gemini that you can interrupt with your voice and it learns your speech patterns and can turn the camera on and point it at things.

  • They didn't really demo that at all, but they did say that they are taking all of the learnings from Project Astra and putting it into Gemini Live.

  • Now, within the Gemini Assistant, they also rolled out Gems, which is a feature that allows users to create customizable AI assistants for very specific tasks.

  • They also showed like a similar thing off in the workspace environment where you can sort of like create your own employee.

  • Now, the workspace example was a little bit interesting because that AI employee had access to every worker's chat history.

  • So it could sort of verify if a project was on track or how things are moving along, which I thought was a little bit strange.

  • But yeah, other than that, they didn't really go in depth in the whole like voice assistant aspect of things.

  • So we don't really know if the latency is coming down or if the voices are changing, but there are some other smaller features that they showed off on the mobile side, such as offering help with a video you're watching or being able to search within a PDF.

  • And then specifically for Pixel devices, by leveraging Gemini Nano, which all of the processing happens entirely on device, it can actually look at your screen, read your conversations and suggest things to say in that conversation.

  • It can even listen into your phone calls and sort of detect whether if it's a scam or not, which is really wild.

  • I'm going to help you transfer your money to a secure account we've set up for you.

  • Now, finally, the last thing that I'll touch on is all of the generative AI that Google is working on, which is all sort of housed under Google Test Kitchen.

  • So music effects and video effects are completely new.

  • This is their music and video generative AI features.

  • With music effects, you can literally type out any instrument and it'll create a completely new beat and you can even layer on multiple instruments, which is crazy.

  • It's pretty tough actually.

  • Video effects is also completely new and showcases some really impressive work in terms of the physics behind it and the detail.

  • Photo effects is getting even more realistic now with AI generated imagery.

  • And Google also briefly touched on Synth ID, which is basically this tool that embeds these invisible watermarks on anything that is generated by AI so that we as humans are able to identify what works of arts has AI's fingerprints on it.

  • But yeah, that was Google I-O 2024.

  • Google really is going all in on AI and it's a little bit overwhelming as the consumer.

  • There's just so much going on.

  • I think as they're slowly rolling these features out, it's gonna take consumers a bit of time to sort of get used to the new workflow of things.

  • But once they do get used to it, it's gonna be a game changer.

  • Unfortunately for a lot of these things, we will have to wait a couple of weeks or even months to see these things getting rolled out.

  • So definitely get subscribed to see my coverage of those things when they do come out.

  • I will be leaving links down below where you can start to enable a lot of these experimental features as well as explore some of these experimental apps.

  • That's gonna do it for this video.

  • If you enjoyed it, leave a like on the way out and I'll catch you guys next time.

  • Peace.

What's up, Josh here.

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it