Google is the latest company to join the artificial intelligence (AI) game. The company has formally introduced a new AI model dubbed Gemini.
“Today we introduced Gemini, our largest and most capable AI model – and the next step in our journey toward making AI helpful for everyone,” the company wrote in a statement. “Built from the ground up to be multimodal, Gemini can generalize and seamlessly understand, operate across, and combine different types of information, including text, images, audio, video, and code.”
Google says that Gemini will be able to run on devices as small as smartphones and as large as data centers, and will feature three different sizes:
- Gemini Nano: the smallest version of the model, used for tasks on devices like smartphones
- Gemini Pro: this model will allow users to carry out a variety of tasks
- Gemini Ultra: the largest version of the model, used for more complex tasks
Gemini’s capabilities
With open AI becoming more and more popular with consumers, how did Google create Gemini to stand out? And how does it stack up against other AI models?
According to Demis Hassabis, CEO and co-founder of Google DeepMind, it all came down to rigorous testing. Based on their results, Gemini is up to the challenge.
“From natural image, audio, and video understanding to mathematical reasoning, Gemini Ultra’s performance exceeds current state-of-the-art results on 30 of the 32 widely-used academic benchmarks used in large language model research and development,” Hassabis wrote in a blog post for the company.
“With a score of 90%, Gemini Ultra is the first model to outperform human experts on massive multitask language understanding, which uses a combination of 57 subjects, such as math, physics, history, law, medicine, and ethics for testing both world knowledge and problem-solving abilities.”
According to the company, the multimodality is one of the primary functions that makes Gemini stand out among other open AI competitors, as it can work with many digital mediums – not just text.
Availability of Gemini
How will consumers be able to access Gemini? Google has big plans for its AI model.
If you have the Pixel 8 Pro phone from Google, keep your eyes peeled. You will be able to utilize Gemini in a few different ways – many of which will debut in the coming months. Using Gemini Nano, users can expect two new features right away: SmartReply and Summarize in the Recordings app.
It remains to be seen whether other Android devices will get access to the Gemini capabilities, or if Google will keep it strictly to its own devices.
Gemini is also now in use on Bard, Google’s AI chatbot, which utilizes the Pro version of the model. While the update has already given Bard more capabilities than before, Google has plans to upgrade it further. Early next year, Bard Advanced, powered by Gemini Ultra, is set to debut.
For those who have Chrome as their web browser and regularly utilize the search engine, there are plans to incorporate Gemini into these features, too.
Starting on December 13, Gemini Pro will officially be available for developers through Google’s AI Studio or Google Cloud Vertex AI. The initial rollout will only be compatible with English, but Google plans to include more languages, and access to Gemini in more territories, in 2024.
Photo credit: James Manyika, SVP, Research, Technology & Society, Google (Screenshot).
- Sponsored Love: Top 5 Common Pests in Orange And How Experts Eliminate Them Effectively
- Sponsored Love: How Can Irvine Property Management Services Maximize Your Rental Income?
- Sponsored Love: Getting Started With Freeview On Your TV
- Sponsored Love: Choosing The Right Personal Injury Lawyer
- Creative Neon Sign Ideas, Incorporating Neon Into Your Aesthetic
Become a Harlem Insider!
By submitting this form, you are consenting to receive marketing emails from: . You can revoke your consent to receive emails at any time by using the SafeUnsubscribe® link, found at the bottom of every email. Emails are serviced by Constant Contact