On May 14th*, Google I/O is taking place. The Google Keynote will start at 10:00 AM PT, followed by the Developer keynote at 1:30 PM PT. I am creating this topic for speculation on what you believe will be revealed, and then after the event takes place, or while it takes place, discussion about what was released.
May 14th
Yeah Iām really curious to see what all is gonna be shown off!
Thanks for the catch .
If there isnāt something about integrating Gemini into Google/Nest devices I would consider that a huge drop of the ball.
Iāve all but stopped using my āsmartā speakers because itās become absurdly clear how dumb they are.
Really looking forward to Gemini Advanced 1.5 being made available to the public.
I think itās about time!
I personally am hoping for a text-to-video model like Imagen Video but more advanced.
Google AI Studio and possible new features
Weāre now less than 3 hours away from the start of the event, and Iām more hyped than ever, especially after seeing this teaser from yesterday:
Somewhat underwhelmed.
What I liked
- Gemini Flash
- Very promising looking model, compares well to Ultra and Pro, great price/performance propositionāespecially being multi-modal. OpenAI not releasing their
gpt-3.5
with vision model yesterday now looks like a mistake.
- Very promising looking model, compares well to Ultra and Pro, great price/performance propositionāespecially being multi-modal. OpenAI not releasing their
- Gemini Nano
- Interesting use-cases like transcribing records of phone calls, automatically adding calendar events etc.
- Would love to see some benchmarks
- Curious if this model will be available to Devs via API.
- If it is meant to be running on edge-devices, can we expect the model to be open? Weights will invariably leak.
- LearnLM
What I didnāt see but had expected/hoped
- A more powerful model
- Gemini 1.5 Ultra
- Gemini 2.0 Pro
- etc
- Gemini in Nest Mini and Nest Hub devices
- Cāmon Google! This is a transformative game changer and a huge ball-drop. Is it a price issue? Please give us a home assistant we actually want to use now!
Lots of features were announced, but a lot of it were things developers are interested in building themselves .
I like the competition announcement though!
Also, the amount of tools and features that are on 20 different platforms is confusing, and I think I already lost track of what got announced with what.
So, are the 1.5 pro models going to be able to natively accept multimodal input and produce multimodal output now in a way that we can leverage with the API, or what?
EDIT: Also forgot to mention, I am excited about the integration with WebGPU
My favorite takeaways from this was alphafold 3 and gameface.
The Alphafold model will be a huge step towards curing diseaseās and save lives. This is a major step towards actually curing Alzheimerās, Parkinsonās, Huntingtonās Disease, Amyotrophic Lateral Sclerosis (ALS), Prion Diseases, Cystic Fibrosis, and Type 2 Diabetes.
And gameface, the AI that lets you control a PC using your face, is a massive step towards getting rid of the āmouth stickā that is commonly used by people who are paralyzed
This will save lives, and Iām happy to be along for the ride!