Introducing Gemini 1.5 Pro: The Next-Generation LLM on Vertex AI

Introducing Gemini 1.5 Pro: The Next-Generation LLM on Vertex AI

In February 2024, Google announced the Gemini 1.5 Pro model, a significant leap from its predecessor, the Gemini 1.0 Ultra. Now available for testing on Vertex AI, this next-generation model is designed to offer enhanced performance and efficiency.

Model Characteristics

The Gemini 1.5 Pro stands out with its new Mixture-of-Experts (MoE) architecture, which ensures better performance while being more efficient to train and serve. One of the key highlights is its extensive context window, which supports up to 128,000 tokens natively and can extend up to 1 million tokens for early testers and eventually all users.

This model is also optimized for multimodal capabilities, handling text, images, and videos seamlessly across a wide range of tasks.

Access and Testing

The Gemini 1.5 Pro was initially released as an experimental model on Vertex AI, accessible via the UI without API support. It is now generally available, with features like context caching and provisioned throughput enhancing its usability.

Features and Capabilities

One of the standout features of the Gemini 1.5 Pro is its long-context understanding, enabling it to process extensive data efficiently. The model also supports code execution, allowing developers to run Python code directly within the model through the Gemini API or Google AI Studio. Additionally, context caching improves speed and reduces costs for AI requests involving repetitive content.

Ethics and Safety

Google places a high emphasis on ethics and safety, and the Gemini 1.5 Pro undergoes extensive testing to align with Google's AI Principles and robust safety policies.

Pricing and Quotas

The pricing structure for the Gemini 1.5 Pro includes tiers starting at the standard 128,000 context window, with options to scale up to 1 million tokens. The experimental model comes with a quota of 2 queries per minute (QPM), which cannot be increased.

Future Developments

Looking ahead, Google is focused on optimizing the model to improve latency, reduce computational requirements, and enhance the overall user experience. Additional features and improvements are on the horizon, including better grounding for accuracy and deeper integration with other Google services.

These advancements make the Gemini 1.5 Pro a powerful tool for developers and enterprises, offering unparalleled capabilities in the realm of AI and machine learning.

Read more