Affiliate links on Android Authority may earn us a commission. Learn more.
Google did the impossible at I/O: It made me feel less apprehensive about AI
Published onMay 14, 2023

βAI is taking over the world, but Iβm not ready for it,β is a sentiment I shared with you a few weeks ago. My apprehensions about generative AI havenβt changed since then, but as I heard the word βAIβ being uttered for the 100th time during this weekβs Google I/O, a quick realization dawned on me: I was OK with that. Maybe not as ecstatic as the hundreds of I/O attendees clapping and laughing like a β90s sitcom track, but still, relatively OK and comfortable with Googleβs approach to generative AI.
Iβve since rewatched the conference once more while rewinding specific moments many times over to analyze them. Iβve also given myself a couple of days to re-assess whether my rose-tinted Google glasses were affecting my judgment here. The result is Iβm still generally comfortable with Googleβs AI approach. Where OpenAIβs GPT4 presentation and Microsoftβs CoPilot launch left me in awe but also intimidated and concerned, Googleβs take seems smartly selective and restrained.
The βresponsibleβ AI message, which was hammered home, resonated with me. I am fully aware that Google came late to the generative AI table and that this concern for doing things right could be β and probably is β a perfect marketing spin to justify its delay and ease investors (who are digging this, by the way; Googleβs stock rose from around $108 earlier this week to $118 at the time of writing and is at its highest since the summer of 2022), all while playing the good-guy role and getting the general public on its side.
I also realize that in an alternate universe, had Google pushed on with generative AI before ChatGPTβs rise to fame, the roles might be reversed and it could be the one bulldozing through without any concern for responsibility or ethics, while other companies urged it to slow down. But we live in this reality and, maybe for the good of humanity as a whole, the biggest web company in the world has adopted* a responsible stance on AI and generative AI**.
*For now.
**In public.
What do you think of Google's generative AI approach so far?
Google crafted a perfectly-balanced message about AI

The screenshot above shows 15 Google products that are actively used by more than half a billion β with a B β people and businesses. Of those, six products serve more than two billion. Google can impact a quarter of the worldβs population in an instant, so imagine if that impact was nefarious.
Thatβs why I am a little admirative of how well crafted the I/O presentation and the message around responsible AI were. Throughout the two hours, Google maintained a masterful balance between:
- we know we have the smarts and power,
- weβre aware this is scary,
- but look, it can also be fun!
- and incredibly useful,
- and weβll hold your hand because we know itβs super intimidating to get started,
- and weβll do our best to keep encouraging original web creators (because we rely on their data to farm it for our answers and ads, dβoh),
- though, you understand, we kinda have to compete with them and write good answers to searches, because otherwise, our product will appear useless to people,
- and yo, investors, we can do so much more, weβre just holding ourselves back now π, but hereβs a small preview of more impressive things to come!
Honestly, for the biggest web company in the world with such a weight on its shoulders, so many different actors to please, and so much tension to navigate, I think Google did the impossible here. Or maybe PaLM 2 did? Who knows who crafted that entire keynote at this point?
Google and generative AI: First, lure them in with familiarity and fun

As a regular web user, Google first eased me in with the innocuously fun and hand-holding aspects of its generative AI approach.
Features like Help me write in Gmail and Google Docs, Help me organize in Sheets, and Help me visualize in Slides put generative AI in a familiar context with clear directives. Unlike ChatGPT, Google Bard, or Bing Chat, youβre not staring at a blank canvas with no idea where to go. You have a springboard from which to launch your ideas while the AI has some boundaries to restrict itself to. There are even clear directives for personalizing the result to make generated texts shorter or more formal or change the styles of images, for example. Sure, that limits creativity and what you can do with this aspect of Googleβs AI application, but Iβm fine with that.
Instead of giving AI a blank canvas, Google has restricted it in many of its popular products.
Other fun implementations, like generative and emoji wallpapers and Magic Compose in Messages, are great, low-stakes ways for people to familiarize themselves with AI-generated content β and for Google to vastly enhance its algorithms and models, letβs be pragmatic.

It was also easier for me to understand the usefulness and power of this kind of AI because Google applied it to products I already use every day. The new immersive View for Routes in Maps sounds incredibly useful when Iβm planning trips to an unknown route or city, Project Tailwind would be an immense help for synthesizing information from dozens of my documents in Google Drive, and synthesized responses in Google Search should understand multi-criterion searches and save me hours of frantic browsing and manual comparisons.
Iβm sure developers appreciated all the generative AI features made specifically for them too.
In many of these instances, the message was that AI-generated content and information are a good starting point, not a final product. Thatβs better than pretending AI in its current state can fully replace a writer, teacher, developer, doctor, or lawyer, when all itβs doing is collecting and organizing the information we humans have provided first. The only difference? Itβs the fastest and most organized assistant ever.
Googleβs AI approach is a little late, but also a little right

Besides the more obvious use cases, Google put forward less controversial applications of AI like helping developers with code writing and testing, and training fine-tuned datasets for threat analysis (Sec-PaLM) and medical research (Med-PaLM). Itβs definitely easier for me to accept and perhaps embrace a predictive AI when I know it can help researchers build a more effective delivery method for cancer treatment.
Plus, there were many hints throughout the conference that showed Google has been β obviously, duh β watching the AI space closely, taking notes of pernicious issues, and thinking ahead about its responsibility as a web leader to do things right and how it can impact the future of AI positively.
AI applications in medicine, fake image watermarking, proper sourcing, and other measures are all good signs for the future of AI.
For example, if embraced by the rest of the AI players, watermarking and metadata for synthetic images should partly reign in misinformation and the dissemination of forged media. And fakes will be highlighted in Google Imagesβ About this image tags or through Google Lens detection.
Proper sourcing, be it from the web for searches, GitHub for coding, or your own documents and emails in Project Tailwind and Sidekick, is a good start. Is it enough to keep web creators well-fed and motivated as well as completely push misinformation away? I highly doubt it, but at least Iβd be able to trace down many answers to their origin instead of blindly trusting it as I do with ChatGPT and other AI models now.
I was also pleasantly surprised to see Google mention its concerns of bias in AI, the importance of nailing different languages correctly before rolling them out willy-nilly, and its automated adversarial testing to reduce inaccuracies and prepare the AI for the million levels of hell that us humans will put it through.
Remember the early days of Bing Chat? Well, itβs better to be as ready as possible for all kinds of similar scenarios when two billion people get unfettered access to generative AI through Google Search.
So while Google is a few months late to the party, it has brought all its advantages to the ring: more users, more implementations, more investments, more developers, and more righteousness. As I mentioned earlier in the article, Iβm glad we live in this version of reality, because if all of those had been used without the βresponsibleβ ethos hanging above them, well,β¦ I donβt know. And Iβm glad we donβt have to find out.
Being late doesn't matter. The moment Google gets in the race, it has a headstart on everyone else.
The scary part is that Google doesnβt have a public headstart yet. As a matter of fact, most of its AI products arenβt live yet or are still in limited previews. But when it gets going, none of that will matter. Itβll be as if Iβd been running a marathon at my snailβs pace of 5K per hour for six whole hours, then Eliud Kipchoge starts his race and catches up to me before I am anywhere near the finish line. With its ubiquity in our everyday lives, Google can catch up to the best players in seconds.
Its data will only get better with time because the more people use it β and there will be two billion right there in Search β the more accurate and nuanced itβll get. AI needs training data and who has data? Google. Who can collect more data? Google.
Game over.
Exercising restrain while pushing the AI envelope

Youβve probably guessed from my writing so far that Iβm teetering on the edge between naive fascination, resignation, excitement, and nervousness. But Iβm still, at least for now, more at ease with Googleβs approach to AI than OpenAI, Microsoft, Midjourney, or other players. Why?
I often got the sense that OpenAI and Microsoftβs approaches, for example, have been so fascinating, but also so rushed, chaotic, overwhelming, and scary. Thereβs a βtech broβ mentality of pushing fast, breaking things, and iterating faster, with complete disregard for consequences. Iβll leave this snippet of Michael Schwarz, Microsoftβs Chief Economist, for you to judge. Me? Iβm thankful Microsoft doesnβt rule the world.
Donβt mistake this for blind trust in Google or hatred for competition β quite the contrary. I am all in for diversification and I think we need a bit of tug and pull, but having someone steer the ship and set a good standard is crucial.
And yes, Iβm sure there are people at Google who think the same way as Michael, but Iβm thankful thatβs not the companyβs public stance. Google appears to at least be asking itself, βWe can, but should we?β And, βIf so, how do we?β
Where OpenAI and Microsoft are moving fast and breaking things, Google appears to be exercising some restrain.
We could all glimpse signs of this restrain throughout I/O. For instance, Google is restricting its AIβs access where it could have the most impact, and giving it more freedom in products you have to actively seek. So while many of the new AI-based features will be siloed into specific popular Google products and use cases, users who donβt want to stifle their creativity or limit themselves can always go to the blank canvas of Bard and its PaLM 2 model. They can use its API and add its integrations to color outside the lines and build solutions tailor-made for them.
Then there were the teasers of more powerful things to come. The next-level Gemini model, Sidekick prompt-generator in Google Workspace, and Universal Translator were all expertly touted to appease investors then swiftly put back in the data centerβs proverbial drawer until theyβre ready and reliable. Another company with a tech-first approach would release some, if not all, of these as soon as possible and shrug the consequences off. Iβm glad Google didnβt.
I lied, I still have apprehensions

My first and main question is: Where is the money?
We know ads are Googleβs bread and butter and we saw glimpses of those through the sponsored product placement in Search but what else? This canβt possibly be the only monetization option. One generative AI query costs more processing power β and thus more money β than a simple search. So how is Google recouping that?
There are bound to be ads and sponsored results above the fold for each search. Does that mean those who pay get preferential treatment among the featured links? Will we find ads and recommended products in Googleβs other AI implementations? I donβt want to see a calculator ad if I ask the AI in Sheets for help with a function!
Then there are questions that could alter the state of the web further, forever. Excellent sites and sources are already dying, if not dead. Good human web writers and journalists are becoming commodities for many publishers; their voices arenβt as heard and their work isnβt as appreciated. So will sites with excellent SEO optimization and poor content still win and get their results on top? Will AI-generated web content create a self-validating loop for featured snippets?
I'm very skeptical of AI monetization, ads, SEO gaming, and what those can do to the future of the web.
Imagine if one AI wrote web content, while another AI read it, parsed it, then featured it at the top of search results. Thatβd be a hell of a future. No humans needed. Where would the original data come from? Pure fabrication?
As a web writer and content creator on a site that lives and dies by its SEO and arbitrary Google rankings, I am also a million times worried about my careerβs future. I canβt do anything about it, though, can I? The train has already left the station. I tell myself that, for now, I just have to focus on creating engaging and personal content (like this, hopefully?) and hope thatβs enough to keep my voice above the fold or, preferably, in Googleβs featured generative AI snippets.
From emoji wallpapers to altered realities, thereβs a single step

Google started I/O 2023 with an impressive and fun demo of the upcoming Magic Editor in Photos. That set the tone for what was to come: a deluge of fascinating generative AI features that straddle the line between enhancement and fabrication, pushing the boundary of reality versus fiction and forcing me to move my own AI tolerance line a few notches ahead.
Announcement after announcement, it dawned on me that we canβt put the genie back in the bottle now. But if the genie has to be out, there are mostly right and mostly wrong ways to do it. For now, Google wins a skeptical nod of approval from me; this is the mostly right way I can get behind.
For now, Google wins a skeptical nod of approval from me, but I remain cautious of everything it chose not to say.
I remain cautious of everything Google chose not to say, though. For instance, framing image generation under the guise of fun wallpapers and innocuous Slides illustrations is a genius way to abolish concerns of fabricated realities and keep the AI model grounded in strict parameters. After all, an emoji wallpaper is a million times less scary than a fake but realistic-style βphotographβ on Midjourney or other image generators. Many of us would be more willing to make a poop and ducks emoji wallpaper than put the Pope in a Balenciaga puffer jacket.
But is Googleβs restrain a temporary illusion to avoid a million pitchforks at once? And will it watch us while we start embracing these fancy new AI features and lose track of our apprehensions, then start slowly pushing those boundaries further and further under the pretense of innovation? In other words, have we learned anything from Facebook?