[ad_1]
With the release of Gemini, Google announced its foundational model for processing a wide range of data sources, including input creation capabilities. This is the company’s most advanced multi-modal model, optimized in three sizes: Ultra, Pro and Nano. The company has been catching up with ChatGPT and other associated OpenAI models since the Chatbot was introduced.
Gemini is reportedly trained on five times more data than Google’s PaLM 2 LLMs, but it has also been upgraded to an internal Tensor Soc and TPU v5p for efficiency and better performance. This makes it faster and reduces the cost of using such a system. With the release of Gemini, OpenAI may try to prove that its GPT-4 or its technology remains smarter than Google’s.
What is Gemini AI?
This multimodal model would be more powerful than GPT 3.5, but also overtakes GPT-4, as reported. Users can use the company’s multimodal model through Bard and the Google Pixel 8 Pro in more than 170 countries.
The company has used open web resources to train Gemini, similar to other GenAI models. This helps with customer service engagement and productivity apps like summarizing meetings, generating code useful to developers, and helping students and researchers find information.
Twin Nano
It is optimized for mobile, enabling features such as smart replies even in E2EE apps like WhatsApp, summaries in the Recorder app, and features such as grammar correction and proofreading via GBoard and contextually on device. Everything happens on-device, which means less latency, and it even works offline. This is already starting to be rolled out to the Google Pixel 8 Pro.
#TeamPixelwe come with presents!🎁#Pixel8 Pro now uses Gemini Nano which powers AI features such as Summarize in Recorder📝 and Smart Reply in Gboard.💬
But that is not everything! Discover how a new #FeatureDrop makes your Pixel (even older ones) feel like new again: https://t.co/E3xkAYBYozpic.twitter.com/MZtMN48DV9— Made by Google (@madebygoogle) December 6, 2023
Moreover, Video Boost mode helps computational photography adjust color, exposure, stabilization and grain. Applying this will give you rich details and colors in videos, meaning that with Night Sight you can even create detailed timelapse videos in low light.
Expansion with Photo Unblur and similar features would be released in the future to improve user experience.
Gemini Pro
It only includes generated responses in text or code format. It is faster and more accurate than the existing PaLM model used in Bard.
Gemini Nano on #Pixel8 Pro is also starting to power Smart Reply in Gboard, which suggests high-quality replies to messages. It’s available now in English in developer preview in WhatsApp, and coming to more apps next year. pic.twitter.com/Lp9stPxU7s— Google (@Google) December 6, 2023
Twin Ultra
It is said that this will be revealed early next year. It will also be accessible via Google Bard as a significant upgrade. It is said to understand text, images and videos. So far, Google hasn’t used a trained AI model like OpenAI’s DALL-E for images and Whisper for audio. We have yet to see how capable the Ultra will be. It reportedly scores 90% in MMLU and covers 57 subjects including Mathematics, Physics, Law, Medicine and Ethics.
Gemini is also one of the world’s leading basic coding models.
It can understand, explain and generate high-quality code in programming languages such as Python, Java, C++ and Go. https://t.co/mwHZTDTBuGpic.twitter.com/MseIT42CIB— Google DeepMind (@GoogleDeepMind) December 6, 2023
In addition, Ultra also updates its code writing tool by introducing AlphaCode 2, which outperforms competitive human computer programmers by 85%. Moreover, it is the first model to outperform human experts in MMLU (Massive Multitask Language Understanding).
Following this, Google also said that this is the first AI model to outperform human experts at specific tasks, and it also surpasses OpenAI’s GPT-4 model in multiple tests.
Google also provides tools and APIs for Android developers to build apps with Gemini. Additionally, if developers want to use the Gemini AI model and AICore, they can participate in the Early Access Program.
Google Gemini AI login
Users can start using Gemini Pro through Bard. However, it is currently out of stock in the UK (it is not being released in the European Economic Area, which includes the EU and Switzerland).
- Visit Google Bard.
- Sign in with your personal Google account.
- Then use Google’s latest model, the Gemini.
- Write a prompt and have content generated for you.
Users can use it with Bard for advanced reasoning, planning, understanding, and other capabilities. However, the company has yet to release its training data, which mainly consists of open web texts such as Wikipedia, Reddit, etc. These are currently available in English but will eventually diversify into other languages. Next year, with the launch of Gemini Ultra to power the Bard Advance, it will help recognize and understand text, photos and videos.
Conclusion
Google also plans to license Gemini through Google Cloud to developers to integrate AI into their applications to improve the experience and generate revenue. Not to mention, Google will also include this in their products and services such as Search, Bard, Chrome, Due AI, and other applications.
With the introduction of Gemini AI, Google is going after its competitor’s main business, which will reportedly be unveiled with iOS and Siri sometime next year. We’ve seen Apple leverage AI to extend predictive text recommendations with the ISO 17 iPhone keyboard.
Due to growing competition in the industry, such as the integration of Microsoft’s OpenAI into Bing Chat, which is powered by GPT-4 Turbo alongside DALL-E 3 and Antropic Claude 2.1 to analyze 150,000 words in a single prompt, Google has not specified any regulatory issues for the delay in Bard’s inaccessibility. Experts have raised concerns about the development of AI as it can perform a range of tasks at human or superhuman levels of intelligence.