The elemental distinction between Microsoft and Google’s dueling mid-Might keynotes this 12 months comes down sizzle versus steak. And that isn’t simply my opinion; I may even quantify the disparity that others apparently additionally ascertained. As famous in my current protection of Microsoft’s 2025 Construct convention, the total keynote ran for a minute (and some seconds) shy of two hours:
However The Verge was in a position to condense the necessities right down to a 15-minutes (and some seconds) abstract video, 1/8th the size of the unique:
What about Google’s day-later different? It was solely a few minutes shorter in whole:
However this time, The Verge was solely in a position to shrink it right down to round 1/3 the unique size, leading to a 32-minute (and alter) abstract video:
Translation: almost the identical keynote length, however rather more “meat” within the Google keynote case. And that’s not even counting the 70-minute developer-tailored keynote that adopted it:
That mentioned, in equity, I’ll level out that Google’s personal abstract video for the keynote was solely 10 minutes lengthy, so…
What did Google’s presenters cowl in these 3+ two-keynote hours, and extra usually throughout the two-day occasion (and its virtual-event precursor)? Glad you requested. Within the sections that observe, I’ll contact on what I believed have been no less than a number of the excessive factors. For extra, try Google’s abstract blogs for the developer group and the public at giant, together with the convention protection abstract pages from of us like 9to5Google, Engadget, The Verge and Wired.
Android (and its variants)
Conceptually much like what Microsoft had carried out, Google determined to launch a few of its information forward of the principle occasion. This time, although, it was one week prior, not two. And the main focus this time was on software program, not {hardware}. Particularly, Google mentioned its upcoming Expressive Design revamp of the core Android UI and related apps, together with deliberate added-and-enhanced options for the O/S and apps, and associated evolutions of the Android variants tailor-made for sensible watches (Put on OS), sensible glasses and headsets (Android XR), autos (Android Auto), shows (Google TV), and some other O/S “spins” I might need missed in the intervening time. Within the course of, Google obtained the bounce on Apple, who will reportedly announce a conceptually comparable revamp for its varied O/Ss in a couple of weeks (keep tuned for my protection)!
I’ll speak extra about Android XR and its related {hardware}, as Google did at I/O itself, in a separate topic-focused part to come back later on this piece.
Multimodal giant language fashions
Gemini, as I’ve mentioned in previous years’ Google I/O stories and different writeups, is the corporate’s suite of proprietary deep studying fashions, all turning into more and more multimodal of their supported information input-and-output variety. There are presently three major variants:
- Professional: For coding and sophisticated prompts
- Flash: For quick efficiency on advanced duties, and
- Flash-lite: For cost-efficient efficiency
Plus, there’s Gemma, a associated set of fashions, this time open supply, which, because of their comparatively low useful resource calls for, are additionally helpful for on-device inference with edge methods.
Newest v2.5 of Gemini Professional and Gemini Flash had each already been unveiled, however at I/O Google touted iterative updates to each of them, bettering responsiveness, accuracy and different metrics. Additionally unveiled, this time first-time, was Gemma 3n, particularly tailor-made for cell gadgets. And in addition newly introduced was Gemini Stay, which helps the real-time evaluation and interpretation of (and response to) dwell audio and video feeds coming from a digital camera and microphone. Should you’re considering this sounds quite a bit like Undertaking Astra, which I discussed on the tail-end of final 12 months’s Google I/O protection (albeit not by identify)…properly, you’d be spot-on.
AI integration into different Google services and products…together with search
Simply as Microsoft is doing with its working system and purposes, Google shouldn’t be solely creating consumer direct-access capabilities to Gemini and Gemma by way of devoted apps and internet interfaces, it’s additionally embedding this core AI intelligence into its different merchandise, reminiscent of Gmail, varied Workspace apps, and Google Drive.
Probably the most important augmentation, in fact, is that of the Google Search engine. It was Google’s first product and stays a dominant income and revenue for it and mother or father firm Alphabet, by advantage of the varied types of paid promoting it associates with search outcomes. You will have already observed the “AI Overview” part that for some time now has appeared on the high of search outcomes pages, containing a abstract clarification of the searched-for matter together with hyperlinks to the pages used to generate that clarification:
Properly, now (as I used to be penning this piece, the truth is!) “AI Mode” has its personal tab on the outcomes web page:
And equally, there’s now an “AI Mode” button on the Google Search dwelling web page:
Google is even testing whether or not to relocate that button to a place the place it might utterly change the longstanding “I’m Feeling Fortunate” button.
It wasn’t too way back when varied tech pundits (current firm excluded, to be clear) have been confidently forecasting the demise of Google’s search enterprise by the hands of upstarts like OpenAI (extra on them later). However the firm’s “deft pivot” to AI teased within the title of this piece has ensured in any other case (no less than till regulatory entities could say in any other case)…maybe too a lot, it seems. As I’ve more and more used AI Overview (now AI Mode), I discover that its search outcomes summaries are sometimes enough to reply my query with out compelling me to click on via to a content-source web page, a non-action (versus custom) that suppresses visitors to that web page. Google has at all times “scraped” web sites to assemble and prioritize search outcomes for a given key phrase or phrase, however by presenting the pages’ data itself, the corporate is now drawing the ire of publishers who’re accusing it of content material theft.
Wealthy content material era
Take generative AI past LLMs (giant language fashions) with their rudimentary enter and output choices (no less than these days, seemingly…simply a few years in the past, I used to be extra sanguine about them!), and also you’re now within the realm of producing practical nonetheless photographs, movies, audio (together with synthesized music) and the like. That is the realm of Google’s Imagen (already at v4), Veo (now v3), and Lyria (v2 and new RealTime) fashions and related merchandise. Veo 3, for instance, kicked off the 2025 Google I/O by way of this spectacular albeit fanciful clip:
Right here’s one other (much less foolish total due to this fact, I’d argue, much more spectacular) one from Google:
Extra synthesized video examples and their related textual content prompts will be discovered on the Veo web page on the Google DeepMind web site. Veo 3 is already in public launch, with oft-impressive albeit typically disturbing outcomes and even real-life mimickers. And mix audio, video and nonetheless photographs, add some extra scripting smarts, and also you’ve obtained the brand new AI filmmaking instrument Circulate:
Who would have thought, just some brief years in the past, that the following Spielberg, Scorsese, Hitchcock, Kubrick, Coppola or [insert your favorite director here] would solely leverage a keyboard and an inference processor cloud cluster as his or her content-creation toolbox? We will not be there but, however we’re getting shut…
Coding assistants
Coding is inventive, too…proper, programmers? Jules is Google’s new asynchronous coding agent, unveiled in Google Labs final December and now in public beta, the place it goes up in opposition to
OpenAI’s lately delivered one-two punch of the internally developed Codex and acquisition (for $3B!) of Windsurf. That mentioned, as VentureBeat additionally notes, it’s not even the one AI-powered coding instrument in Google’s personal arsenal: “Google provides Code Help, AI Studio, Jules and Firebase”.
Android XR-based merchandise (and partnerships)
Google co-founder Sergey Brin made a curious onstage confession throughout a “hearth chat” session at Google I/O, admitting that he “made numerous errors with Google Glass”:
His critique of himself and the corporate he led was predominantly two-fold in nature:
- Google tried to “go it alone” from a {hardware} growth, manufacturing and advertising and marketing standpoint, versus partnering with a longtime glasses provider reminiscent of Italian eyewear firm Luxottica, with whom Meta has co-developed two generations (thus far) of sensible glasses (as you’ll quickly find out about in additional element by way of an upcoming sorta-teardown by yours actually), and
- The bulbous liquid crystal on silicon (LCoS) show in entrance of one of many wearer’s eyes ensured that no person would mistake them for a traditional pair of glasses…a differentiation which was not advantageous for Google.
Judging from the 2025 Google I/O messaging, the corporate appears decided to not make the identical mistake once more. It’s partnering with Warby Parker, Korea-based Light Monster, Samsung and Xreal (and presumably others sooner or later) on sensible glasses primarily based on its Android XR platform…glasses that it hopes of us will truly wish to be seen sporting in public. Samsung can also be Google’s lead companion for a VR headset primarily based on Android XR…the “prolonged actuality” (XR) that Google envisions for the working system spans each sensible glasses—with and with out built-in augmented actuality shows—and head-mounted shows. And it not solely did dwell demos through the keynote but additionally gave attendees the possibility to (briefly) check out its prototype sensible glasses, glimpsed a 12 months in the past within the Undertaking Astra clip I discussed earlier, for themselves.
Google Beam
Two years in the past, I famous that the way-cool Undertaking Starline hologram-based digital conferencing sales space system introduced two years earlier (throughout COVID-19 lockdowns; how apropos):
had subsequently been considerably slimmed down and in any other case simplified:
Quick ahead two extra years to the current and Google has rebranded the 3D-rendering know-how as Beam, in preparation for its productization by companions reminiscent of HP and Zoom:
And within the course of, Google has notably added close to real-time, AI-powered bidirectional language translation to the combination (in addition to to its baseline Google Meet videoconferencing service, which beforehand relied on captions), preserving every speaker’s tone and talking fashion within the course of:
Now there’s a sensible utility for AI that I can enthusiastically get behind!
OpenAI’s predictable (counter)punch
In closing, one ultimate point out of certainly one of Google’s major rivals. Final 12 months, OpenAI tried to proactively upstage Google by asserting ChatGPT’s superior voice mode someday forward of Google I/O. This time, OpenAI tried to suck the wind out of Google’s sails retroactively, by trumpeting that it was shopping for (for $6.5B!) the “io” {hardware} division of Jony Ive’s design studio, LoveFrom, someday after Google I/O. To not point out the $3M allegedly spent on the “odd” (I‘m being charitable right here) video that accompanied the announcement:
Whereas I don’t in any respect low cost OpenAI’s future prospects (or Meta’s, for that matter, or anybody else’s), I additionally don’t low cost Google’s inherent benefit in creating customized AI: it’s guided by the truth that it already is aware of (for higher and/or worse) quite a bit about numerous us.
How do you assume this’ll all play out sooner or later? And what did you concentrate on all of the information and enhanced applied sciences and merchandise that Google lately unveiled? Let me (and your fellow readers) know within the feedback!
—Brian Dipert is the Editor-in-Chief of the Edge AI and Imaginative and prescient Alliance, and a Senior Analyst at BDTI and Editor-in-Chief of InsideDSP, the corporate’s on-line publication.
Associated Content material
- The 2024 Google I/O: It’s (just about) all about AI progress, should you didn’t already guess
- The 2023 Google I/O: It’s all about AI, don’t cha know
- Sensible glasses skepticism: A take a look at their previous, current, and future(?)
- Google’s fall…err…summer time launch: One-upping Apple with a sizeable product tranche
The submit The 2025 Google I/O convention: A deft AI pivot sustains the corporate’s relevance appeared first on EDN.