Google's largest and most capable AI model. Built from the ground up to be multimodal, Gemini can generalize and seamlessly understand, operate across and combine different types of information, including text, images, audio, video and code.
This is great but like it goes with all other Google AI products, it needs to be launched in public and tested by devs in real-time to see how it fares against other top LLMs. Ultra is not going to be released soon and as a developer, I'm mostly reluctant to switch to that unless it's significantly better than GPT-4 because of the slower pace of updates and the potential likeliness of Google killing the model in a couple of years.
Thanks! More SotA models to make this market non monopoly)
Report
Hi guys, congrats on launch, checked out how Gemini vision works on Bark - there is room for improvement. Uploaded a pic of sofa in living room - Gemini detected sofa and door and tv and lots of other stuff there was not on the pic :)
We use Vertext AI in our sustainable travel planner www.biliki.ai . Looking forward to 13th December when the new Gemini model will be added to Vertex. Let's meet a new era together.
Nice 🚀💯
Now we use OpenAI API for our AI Furniture Planner. But it will be very interesting try the same in Gemini.
ps. Today is AI Furniture Planner Launch on PH 👋
Gemini, as Google's answer to GPT-4, seems like a significant advancement in AI technology. Wishing the team success! How does Gemini's multimodal capabilities enhance its performance in understanding and operating across various types of information?
Replies
ShortVideoGen
Kommunicate
Scade.pro
INK'A
SprintsQ