Google is the latest company to join the artificial intelligence (AI) game. The company has formally introduced a new AI model dubbed Gemini.
“Today we introduced Gemini, our largest and most capable AI model – and the next step in our journey toward making AI helpful for everyone,” the company wrote in a statement. “Built from the ground up to be multimodal, Gemini can generalize and seamlessly understand, operate across, and combine different types of information, including text, images, audio, video, and code.”
Google says that Gemini will be able to run on devices as small as smartphones and as large as data centers, and will feature three different sizes:
- Gemini Nano: the smallest version of the model, used for tasks on devices like smartphones
- Gemini Pro: this model will allow users to carry out a variety of tasks
- Gemini Ultra: the largest version of the model, used for more complex tasks
With open AI becoming more and more popular with consumers, how did Google create Gemini to stand out? And how does it stack up against other AI models?
According to Demis Hassabis, CEO and co-founder of Google DeepMind, it all came down to rigorous testing. Based on their results, Gemini is up to the challenge.
“From natural image, audio, and video understanding to mathematical reasoning, Gemini Ultra’s performance exceeds current state-of-the-art results on 30 of the 32 widely-used academic benchmarks used in large language model research and development,” Hassabis wrote in a blog post for the company.
“With a score of 90%, Gemini Ultra is the first model to outperform human experts on massive multitask language understanding, which uses a combination of 57 subjects, such as math, physics, history, law, medicine, and ethics for testing both world knowledge and problem-solving abilities.”
According to the company, the multimodality is one of the primary functions that makes Gemini stand out among other open AI competitors, as it can work with many digital mediums – not just text.
Availability of Gemini
How will consumers be able to access Gemini? Google has big plans for its AI model.
If you have the Pixel 8 Pro phone from Google, keep your eyes peeled. You will be able to utilize Gemini in a few different ways – many of which will debut in the coming months. Using Gemini Nano, users can expect two new features right away: SmartReply and Summarize in the Recordings app.
It remains to be seen whether other Android devices will get access to the Gemini capabilities, or if Google will keep it strictly to its own devices.
Gemini is also now in use on Bard, Google’s AI chatbot, which utilizes the Pro version of the model. While the update has already given Bard more capabilities than before, Google has plans to upgrade it further. Early next year, Bard Advanced, powered by Gemini Ultra, is set to debut.
For those who have Chrome as their web browser and regularly utilize the search engine, there are plans to incorporate Gemini into these features, too.
Starting on December 13, Gemini Pro will officially be available for developers through Google’s AI Studio or Google Cloud Vertex AI. The initial rollout will only be compatible with English, but Google plans to include more languages, and access to Gemini in more territories, in 2024.
Photo credit: James Manyika, SVP, Research, Technology & Society, Google (Screenshot).
- New Book Uncovers The Stunning Stories Behind West Africa’s Hidden Photographic Treasures
- NYC Releases Workers’ Bill Of Rights: Your Labor Rights Guide
- Grubhub And Others Launch $200K Microgrant For NYC BIPOC Restaurateurs
- New York City Public Schools Unveils “Hidden Voices: Stories Of The Global African Diaspora”
- Bronx: Black K.A.R.E. Wellness Health Fair In Co-op City