top of page
  • Kyle Chua

Google Admits Gemini Demo Video Not Conducted in Real Time

Google is drawing flak for tricking the public into thinking its latest artificial intelligence (AI) model is more capable than it actually is.

Google Gemini Demo Video
Credit: Google

The search engine provider last week launched what it touted as its fastest and most flexible AI model yet, Gemini, showcasing its capabilities in a six-minute demo video released to the media and to the public. The video shows how Gemini can respond to spoken-word prompts and videos in real time as though it were in a real conversation. It also shows how the new model can describe a drawing of a duck versus that of a rubber duck, among other examples.


In the video's description, Google includes a disclaimer that reads, "For the purposes of this demo, latency has been reduced, and Gemini outputs have been shortened for brevity." That disclaimer, along with other important information about the Gemini's capabilities, however, are nowhere to be found in the video itself.

As a result, the public thought that the demo in the video was all happening in real time. Google only later admitted to Bloomberg that that wasn't the case. In a statement, the tech giant described the video as "an illustrative depiction of the possibilities of interacting with Gemini, based on real multimodal prompts and outputs from testing", suggesting it doesn't acurately reflect the version of the service that's releasing on 13 December.


It's not the first time Google's AI demos sparked controversy. Some of the search giant's own employees called it out earlier in the year for supposedly rushing out the demo of the chatbot Bard in an attempt to compete with Microsoft's own showing that same week.


Google reportedly scrapped plans to hold an in-person launch of Gemini, likely because it thought a virtual showing, where the AI's response speeds are shortened by editing, could seem more advanced than it actually is.


Still, the new AI model is nonetheles impressive as it's said to be natively multimodal, meaning it can process not only text but also images, videos and audio. In comparison, OpenAI's GPT-4 model, its latest and most advanced model so far, is primarily a text-based model, and only becomes multimodal with the use of plugins.

 
  • Goolge launched what it touted as its fastest and most flexible AI model yet, Gemini, showcasing its capabilities in a six-minute demo video released to the media and to the public.

  • The video, however, tricked the public into thinking the user's interactions with the AI were happening in real time, when in reality they weren't.

  • In a statement, Google said the video was just "an illustrative depiction of the possibilities of interacting with Gemini, based on real multimodal prompts and outputs from testing", suggesting it doesn't acurately reflect the version of the service that's releasing on 13 December.


As technology advances and has a greater impact on our lives than ever before, being informed is the only way to keep up.  Through our product reviews and news articles, we want to be able to aid our readers in doing so. All of our reviews are carefully written, offer unique insights and critiques, and provide trustworthy recommendations. Our news stories are sourced from trustworthy sources, fact-checked by our team, and presented with the help of AI to make them easier to comprehend for our readers. If you notice any errors in our product reviews or news stories, please email us at editorial@tech360.tv.  Your input will be important in ensuring that our articles are accurate for all of our readers.

bottom of page