Gemini is a brand-new type of AI that can do things we could only dream of before. It’s not like other chatbots or algorithms.
Google has indirectly challenged OpenAI ChatGPT with their multimodal Gemini AI, which means Gemini AI has text, audio, and visual modalities.
There is a video that dropped on the Google channel titled “Hands-on with Gemini: Interacting with multimodal AI“
A user and a Gemini AI chatbot can be heard talking in the six-minute video. The video also shows Gemini’s ability to tell the difference between pictures and real things. It was amazing that Gemini could read out loud descriptions of drawings of ducks and tell the difference between a drawing of a duck and a rubber duck, among other things.
To be honest, the demo wasn’t done in real-time or in voice either. Bloomberg Opinion asked Google about the video, and a spokesperson said it was made “using still image frames from the footage and prompting via text.”
The spokesperson also pointed to a site where people could use photos of their hands, drawings, or other items to interact with Gemini.