Gemma 3 from google : Multimodality and up to 128k long context | Llm dataset format | Quick start guide to large language models github | Large language model rankings | Turtles AI

Gemma 3 from google : Multimodality and up to 128k long context
Google Gem 3: An Advanced AI Model for Faster, Safer Applications on a Single GPU
Isabella V12 March 2025

 

Google has launched Gemma 3, its new AI model that stands out for its power, versatility, and ability to run on a single GPU. With advanced capabilities for analyzing text, images, and video, Gemma 3 represents an important step towards the evolution of accessible and scalable AI.

Key points:

  • Exceptional performance on a single GPU and TPU.
  • Support for over 140 languages, with multimodal capabilities.
  • Expanded context window for up to 128,000 tokens.
  • Introducing ShieldGemma 2 for image application security.


Google recently announced the launch of Gemma 3, a new iteration of its family of AI models that promises higher performance, optimized to run on a single GPU. This new version expands on the capabilities of previous models, introducing support for multimodal analysis that integrates text, images, and short videos, and placing an emphasis on efficiency and accessibility. Intended primarily for developers and researchers, Gemma 3 is designed to run optimally on Nvidia GPUs, but is also compatible with other hardware architectures, including Google TPU processors, thus expanding its use in a wide range of applications. Just twelve months after the debut of the first version, this new model brings with it significant improvements in terms of capacity and performance.

One of the distinctive aspects of Gemma 3 is its ability to support over 140 languages, including the main international languages, but also less common languages, such as those spoken in some Asian and African regions. The introduction of an expanded context window of 128,000 tokens allows the model to process and understand much more complex text, supporting in-depth analysis and more complex operations. Thanks to this extension, Gemma 3 can handle larger contexts, improving the accuracy of its responses and machine learning processes. Furthermore, the integration of advanced image and video analysis capabilities enables the creation of interactive and intelligent applications, capable of recognizing visual and textual content in real time, opening new horizons in the field of AI-based solutions.

One of the most significant features of this new version concerns the optimization of quantized models. Quantized versions reduce computational needs, allowing to maintain high performance while lowering hardware requirements. This represents a significant advantage for developers who want to implement AI solutions on low-power devices or with limited hardware resources, such as laptops and mobile devices. This improvement in resource management makes Gemma 3 an interesting choice for multiple application scenarios, from research to the production of software for the consumer market.

Despite the high degree of openness and accessibility that Gemma 3 offers, licensing restrictions continue to raise debates regarding the definition of an "open model". Google has maintained some limitations on the allowed uses of the model, in particular regarding the most sensitive or dangerous applications. However, the company said it has taken steps to ensure responsible use of Gemma 3, including robust security protocols and measures to protect against harmful content. In particular, it has introduced ShieldGemma 2, an image classification system designed to filter sensitive content, such as violent or sexually explicit material, at the input and output stages of AI applications. This solution guarantees an additional level of control for developers and a safer use of the model in real-world environments.

Gemma 3 also stands out for its integration with a number of tools already widely used in the industry, such as Hugging Face, PyTorch and TensorFlow. Developers can choose from several deployment platforms, including Google Vertex AI and Cloud Run, to implement scalable and easily integrated AI solutions. In addition, support for the NVIDIA framework allows users to obtain maximum performance even on smaller GPUs, such as those found in embedded devices. With the growing number of community-developed models and custom tools, Gemma 3 fits into an ever-expanding ecosystem that stimulates innovation and knowledge sharing.

The availability of Google Cloud credits offered through the Gemma 3 Academic program represents a further opportunity for researchers to explore the potential of this model in the scientific field.

Thanks to easy distribution through platforms such as Kaggle and Ollama, and the possibility of fine-tuning on their datasets, Gemma 3 is configured as a tool

Video