Launched in early 2024, Gemma is Google DeepMind’s family of lightweight open-source large language models (LLMs). Unlike its larger sibling Gemini, which powers Google’s Bard chatbot and other premium services, Gemma is built for accessibility, flexibility, and research. The name “Gemma” comes from the Latin word for “precious stone,” underscoring its role as a polished, compact version of Google’s broader AI ambitions.
What is Gemma AI?
Gemma is a collection of compact, open-weight generative AI models designed to be easily integrated into your apps, run on your own hardware or mobile devices, and deployed across cloud services. Built on the same foundation as Google’s Gemini models, Gemma is tailored for developers and researchers who want adaptable, fine-tunable AI tools they can shape to suit specific needs.
These models are fully customisable—thanks to their open weights—allowing you to fine-tune them using your preferred AI framework and tools like the Vertex AI SDK. Whether you’re optimising for a particular use case or training the model to better understand your data, Gemma makes it straightforward.
What sets Gemma apart?
Gemma is designed to be small enough to run on local hardware but powerful enough to perform tasks like coding, translation, summarisation, and dialogue. It’s available in 2B and 7B parameter sizes, optimized for speed and efficiency. These models can be run on laptops, desktops, or single-cloud GPUs—without requiring the infrastructure of a supercomputer cluster.
Think of it as Google’s answer to Meta’s LLaMA or Mistral’s models: lightweight, fine-tunable, and open-weight, meaning researchers and developers can tinker under the hood without license restrictions. It’s trained with techniques borrowed from Gemini’s development, including reinforcement learning and safety tuning, but packaged for broader experimentation.
Open but responsible
Google released Gemma with a strong emphasis on “responsible AI.” Alongside the models, it published detailed documentation, model cards, and a Responsible Generative AI Toolkit to guide ethical usage. While the models are open-weight, the license still restricts potentially harmful use cases—something Google sees as a compromise between openness and safety.
Who is Gemma for?
Gemma targets AI researchers, startups, and developers who want cutting-edge performance without being locked into proprietary APIs. It’s optimized for frameworks like PyTorch, TensorFlow, and JAX, and supported out of the box by platforms like Hugging Face, Kaggle, and Colab.
This makes it ideal for anyone looking to:
•Build private or on-device AI tools
•Experiment with model fine-tuning
•Develop language-based applications without cloud dependency
What’s next?
Gemma is only the first phase. Google has hinted at expanding the Gemma family with multimodal and larger models—some expected to support audio and image inputs. These could serve as a bridge between research-grade LLMs and Gemini’s enterprise offerings.
Discover the latest Business News, Sensex, and Nifty updates. Obtain Personal Finance insights, tax queries, and expert opinions on Moneycontrol or download the Moneycontrol App to stay updated!