![]() Optimized for Google Cloud: Vertex AI provides a broad MLOps toolset with a range of tuning options and one-click deployment using built-in inference optimizations.Cutting-edge hardware platforms: We’ve partnered with NVIDIA to optimize Gemma for NVIDIA GPUs, from data center to the cloud to local RTX AI PCs, ensuring industry-leading performance and integration with cutting-edge technology.Cross-device compatibility: Gemma models run across popular device types, including laptop, desktop, IoT, mobile and cloud, enabling broadly accessible AI capabilities.Multi-framework tools: Bring your favorite framework, with reference implementations for inference and fine-tuning across multi-framework Keras 3.0, native PyTorch, JAX, and Hugging Face Transformers.Gemma supports a wide variety of tools and systems: You can fine-tune Gemma models on your own data to adapt to specific application needs, such as summarization or retrieval-augmented generation (RAG). Optimized across frameworks, tools and hardware Guidance: You can access best practices for model builders based on Google’s experience in developing and deploying large language models.Debugging: A model debugging tool helps you investigate Gemma's behavior and address potential issues. ![]() Safety classification: We provide a novel methodology for building robust safety classifiers with minimal examples.We’re also releasing a new Responsible Generative AI Toolkit together with Gemma to help developers and researchers prioritize building safe and responsible AI applications. These evaluations are outlined in our Model Card. To understand and reduce the risk profile for Gemma models, we conducted robust evaluations including manual red-teaming, automated adversarial testing, and assessments of model capabilities for dangerous activities. Additionally, we used extensive fine-tuning and reinforcement learning from human feedback (RLHF) to align our instruction-tuned models with responsible behaviors. As part of making Gemma pre-trained models safe and reliable, we used automated techniques to filter out certain personal information and other sensitive data from training sets. Gemma is designed with our AI Principles at the forefront. See the technical report for details on performance, dataset composition, and modeling methodologies. Notably, Gemma surpasses significantly larger models on key benchmarks while adhering to our rigorous standards for safe and responsible outputs. And Gemma models are capable of running directly on a developer laptop or desktop computer. This enables Gemma 2B and 7B to achieve best-in-class performance for their sizes compared to other open models. Gemma models share technical and infrastructure components with Gemini, our largest and most capable AI model widely available today. Terms of use permit responsible commercial usage and distribution for all organizations, regardless of size.Optimization across multiple AI hardware platforms ensures industry-leading performance, including NVIDIA GPUs and Google Cloud TPUs.Pre-trained and instruction-tuned Gemma models can run on your laptop, workstation, or Google Cloud with easy deployment on Vertex AI and Google Kubernetes Engine (GKE). ![]() Ready-to-use Colab and Kaggle notebooks, alongside integration with popular tools such as Hugging Face, MaxText, NVIDIA NeMo and TensorRT-LLM, make it easy to get started with Gemma.We’re providing toolchains for inference and supervised fine-tuning (SFT) across all major frameworks: JAX, PyTorch, and TensorFlow through native Keras 3.0.A new Responsible Generative AI Toolkit provides guidance and essential tools for creating safer AI applications with Gemma.Each size is released with pre-trained and instruction-tuned variants. We’re releasing model weights in two sizes: Gemma 2B and Gemma 7B.Gemma is available worldwide, starting today. Developed by Google DeepMind and other teams across Google, Gemma is inspired by Gemini, and the name reflects the Latin gemma, meaning “precious stone.” Accompanying our model weights, we’re also releasing tools to support developer innovation, foster collaboration, and guide responsible use of Gemma models. Gemma is a family of lightweight, state-of-the-art open models built from the same research and technology used to create the Gemini models. Today, we’re excited to introduce a new generation of open models from Google to assist developers and researchers in building AI responsibly. We have a long history of contributing innovations to the open community, such as with Transformers, TensorFlow, BERT, T5, JAX, AlphaFold, and AlphaCode. ![]() At Google, we believe in making AI helpful for everyone.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |