
Posted by Tris Warkentin – Director, Product Administration and Jane Superb – Senior Product Supervisor

In February we introduced Gemma, our household of light-weight, state-of-the-art open fashions constructed from the identical analysis and expertise used to create the Gemini fashions. The group’s unbelievable response – together with spectacular fine-tuned variants, Kaggle notebooks, integration into instruments and companies, recipes for RAG utilizing databases like MongoDB, and much extra – has been actually inspiring.
At this time, we’re excited to announce our first spherical of additives to the Gemma household, increasing the chances for ML builders to innovate responsibly: CodeGemma for code completion and technology duties in addition to instruction following, and RecurrentGemma, an efficiency-optimized structure for analysis experimentation. Plus, we’re sharing some updates to Gemma and our phrases aimed toward enhancements primarily based on invaluable suggestions we have heard from the group and our companions.
Introducing the primary two Gemma variants
CodeGemma: Code completion, technology, and chat for builders and companies
Harnessing the inspiration of our Gemma fashions, CodeGemma brings highly effective but light-weight coding capabilities to the group. CodeGemma fashions can be found as a 7B pretrained variant that makes a speciality of code completion and code technology duties, a 7B instruction-tuned variant for code chat and instruction-following, and a 2B pretrained variant for quick code completion that matches in your native pc. CodeGemma fashions have a number of benefits:
- Clever code completion and technology: Full traces, features, and even generate complete blocks of code – whether or not you are working domestically or leveraging cloud assets.
- Enhanced accuracy: Educated on 500 billion tokens of primarily English language knowledge from internet paperwork, arithmetic, and code, CodeGemma fashions generate code that is not solely extra syntactically appropriate but in addition semantically significant, serving to scale back errors and debugging time.
- Multi-language proficiency: Your invaluable coding assistant for Python, JavaScript, Java, and different common languages.
- Streamlined workflows: Combine a CodeGemma mannequin into your improvement setting to jot down much less boilerplate, and concentrate on attention-grabbing and differentiated code that issues – sooner.
![]() |
| This desk compares the efficiency of CodeGemma with different comparable fashions on each single and multi-line code completion duties. Study extra within the technical report. |
Study extra about CodeGemma in our report or strive it on this quickstart information.
RecurrentGemma: Environment friendly, sooner inference at larger batch sizes for researchers
RecurrentGemma is a technically distinct mannequin that leverages recurrent neural networks and native consideration to enhance reminiscence effectivity. Whereas reaching comparable benchmark rating efficiency to the Gemma 2B mannequin, RecurrentGemma’s distinctive structure leads to a number of benefits:
- Diminished reminiscence utilization: Decrease reminiscence necessities permit for the technology of longer samples on gadgets with restricted reminiscence, resembling single GPUs or CPUs.
- Larger throughput: Due to its decreased reminiscence utilization, RecurrentGemma can carry out inference at considerably larger batch sizes, thus producing considerably extra tokens per second (particularly when producing lengthy sequences).
- Analysis innovation: RecurrentGemma showcases a non-transformer mannequin that achieves excessive efficiency, highlighting developments in deep studying analysis.
![]() |
| This chart reveals how RecurrentGemma maintains its sampling pace no matter sequence size, whereas Transformer-based fashions like Gemma decelerate as sequences get longer. |
To grasp the underlying expertise, take a look at our paper. For sensible exploration, strive the pocket book, which demonstrates find out how to finetune the mannequin.
Constructed upon Gemma foundations, increasing capabilities
Guided by the identical rules of the unique Gemma fashions, the brand new mannequin variants provide:
- Open availability: Encourages innovation and collaboration with its availability to everybody and versatile phrases of use.
- Excessive-performance and environment friendly capabilities: Advances the capabilities of open fashions with code-specific area experience and optimized design for exceptionally quick completion and technology.
- Accountable design: Our dedication to accountable AI helps make sure the fashions ship protected and dependable outcomes.
- Flexibility for various software program and {hardware}:
- Each CodeGemma and RecurrentGemma: Constructed with JAX and appropriate with JAX, PyTorch, , Hugging Face Transformers, and Gemma.cpp. Allow native experimentation and cost-effective deployment throughout numerous {hardware}, together with laptops, desktops, NVIDIA GPUs, and Google Cloud TPUs.
- CodeGemma: Moreover appropriate with Keras, NVIDIA NeMo, TensorRT-LLM, Optimum-NVIDIA, MediaPipe, and availability on Vertex AI.
- RecurrentGemma: Assist for all of the aforementioned merchandise might be out there within the coming weeks.
Gemma 1.1 replace
Alongside the brand new mannequin variants, we’re releasing Gemma 1.1, which incorporates efficiency enhancements. Moreover, we have listened to developer suggestions, mounted bugs, and up to date our phrases to offer extra flexibility.
Get began in the present day
These first Gemma mannequin variants can be found in numerous locations worldwide, beginning in the present day on Kaggle, Hugging Face, and Vertex AI Mannequin Backyard. Here is find out how to get began:
We invite you to strive the CodeGemma and RecurrentGemma fashions and share your suggestions on Kaggle. Collectively, let’s form the way forward for AI-powered content material creation and understanding.


