Google adds two new AI models to its Gemma family of LLMs - Why this is important | Tech News

Google adds two new AI models to its Gemma family of LLMs - Why this is important

Google introduces two new AI models, CodeGemma and RecurrentGemma, promising possibilities for innovation for Machine Learning (ML) developers.

By: SHAURYA TOMER
| Updated on: Apr 11 2024, 12:05 IST
From Google Pixel 8a to Android 15, what to expect at Google I/O 2024
Google adds two new AI models to its Gemma family of LLMs - Why this is important
1/5 Google I/O 2024 time and place: Google has announced that its annual developer conference will kick off on May 14 at 10:30 PM IST or 10:00 AM PT. If last year's trend continues, it is likely to be held at the Shoreline Amphitheater at the Google HQ in Mountain View, California. (Google)
Google adds two new AI models to its Gemma family of LLMs - Why this is important
2/5 1. Artificial intelligence: During this year's conference, artificial intelligence (AI) could yet again take center stage, especially after Google already made several major AI announcements this year including Gemini and Gemma, its families of Large Language Models (LLMs). (Google)
Google adds two new AI models to its Gemma family of LLMs - Why this is important
3/5 2. Google Pixel 8a: Google is also likely to showcase its upcoming affordable smartphone, the Pixel 8a. It will be the latest entry into the Google Pixel 8 series which already has the Pixel 8 and Pixel 8 Pro. In January, an alleged Pixel 8a retail box was spotted online which hints at a Pixel 8-esque design of the rumoured smartphone. (Representative Image) (Shaurya Tomer/HT Tech)
image caption
4/5 Android 15: The Android 14 rollout began last year and more and more devices are getting it. What does this mean? We could potentially see some Android 15-related unveilings or announcements at Google I/0 2024, especially now that the Preview Phase 1 seems to have kicked off. (Unsplash)
image caption
5/5 Other announcements: Apart from this, Google could also make announcements related to its other services including Gmail, Photos, Maps and Workspace. (Unsplash)
Google adds two new AI models to its Gemma family of LLMs - Why this is important
icon View all Images
Know all about the new AI models introduced by Google - CodeGemma and Recurrent Gemma. (Google)

In February, Google took the wraps off Gemma, its family of lightweight Large Language Models (LLMs) for open-source developers. Researchers at Google DeepMind developed it intending to assist developers and researchers in building AI responsibly. It has now announced two new additions to Gemma - CodeGemma and RecurrentGemma. With this move, Google DeepMind aims to keep up the pace in the artificial intelligence (AI) race, facing competition from the likes of OpenAI and Microsoft. 

Also Read: Google Gemini AI images disaster - What really happened with the image generator?

While the company has found itself in hot waters over some of the AI capabilities of its most popular AI model, Gemini, it seems that the controversy has not slowed down researchers. These new AI models promise possibilities for innovation for Machine Learning (ML) developers. Know all about the two new Gemma AI models - CodeGemma and Recurrent Gemma.

Google CodeGemma

The first of the two new AI models is CodeGemma, a lightweight model with coding and instruction following capabilities. It is available in three variants:

1. 7B pre-trained variant for code completion and code generation tasks

2. 7B instruction-tuned variant for instruction following and code chat.

3. 2B pre-trained variant for quick code completion on local PCs.

Google says CodeGemma can not only generate lines, and functions but can even create blocks of code, irrespective of whether it is being used locally on PCs or via cloud resources. It has multi-language proficiency, meaning you can use it as an assistant while coding in languages such as Python, JavaScript and Java. The code generated by CodeGemma is not only advertised as being syntactically accurate but also right semantically. This promises to cut down on errors and debug time. 

Also Read: Know all about Gemma - Google's family of LLMs

This new AI model is trained on 500 billion tokens of data which is primarily English, including code from publicly available repositories, mathematics and documents on the web. 

Google Recurrent Gemma

The other AI model, called RecurrentGemma, aims to improve memory efficiency by leveraging recurrent neural networks and local attention. Thus, it is meant for research experimentation. While it delivers similar benchmark performance to DeepMind's Gemma 2B AI model, RecurrentGemma has a unique architecture that allows it to deliver on three fonts - reduced memory usage, higher throughput and research innovation.

Also Read: Apple in talks with Google over licensing Gemini for AI features on iPhones

As per Google, RecurrentGemma can generate longer samples even on devices with limited memory due to the lower memory requirements. This also allows the AI model to carry out inference in large batches, increasing the tokens per second. Google also says Transformer-based models like Gemma can slow down as sequences get longer. On the other hand, RecurrentGemma maintains its sampling speed irrespective of the sequence length.

Google says it shows a “non-transformer model that achieves high performance, highlighting advancements in deep learning research.”

One more thing! We are now on WhatsApp Channels! Follow us there so you never miss any updates from the world of technology. ‎To follow the HT Tech channel on WhatsApp, click here to join now!

Catch all the Latest Tech News, Mobile News, Laptop News, Gaming news, Wearables News , How To News, also keep up with us on Whatsapp channel,Twitter, Facebook, Google News, and Instagram. For our latest videos, subscribe to our YouTube channel.

First Published Date: 11 Apr, 12:05 IST
NEXT ARTICLE BEGINS