top of page

Google AI Studio: All Models Available, Modality Coverage, Developer Access and Deployment Structure

ree

Google AI Studio provides a unified interface for testing, comparing and integrating Google’s full ecosystem of generative-AI models, including the complete Gemini 2.5 family, preview variants, multimodal systems, speech and audio models, image and video generators, and lightweight open-model alternatives for research and on-device use.

The platform exposes all models through a consistent developer experience, supports programmatic discovery through the Gemini API, and allows real-time experimentation with text, images, audio and video inputs across stable and preview model classes.

Google AI Studio therefore functions as the central hub through which developers, researchers, analysts and enterprise users can explore model capabilities, check token limits, test multimodal prompts and connect models to production environments using Google Cloud or direct API access.

··········

··········

Google AI Studio provides access to the complete Gemini 2.5 family, including Pro, Flash, Flash-Lite and preview models.

The Gemini 2.5 line represents Google’s largest, most capable model suite available in AI Studio, organized into multiple tiers that balance reasoning depth, latency, context size and cost efficiency.

Gemini 2.5 Pro serves as the flagship reasoning model with advanced multimodal capabilities, large context depth, high-precision output and support for deeply structured tasks.

Gemini 2.5 Flash offers balanced performance for general-purpose workloads, maintaining strong multimodal processing and large context capacity with lower operational cost.

Gemini 2.5 Flash-Lite delivers cost-efficient, high-throughput inference suitable for real-time applications, automations, mobile integrations and large-scale backend processing.

AI Studio also exposes preview models such as Flash Live, TTS variants and native audio-dialogue models that support real-time streaming and experimental modalities under controlled access.

·····

Gemini 2.5 Model Availability

Model Tier

Identifier

Capability Profile

2.5 Pro

gemini-2.5-pro

Deep reasoning + multimodal

2.5 Flash

gemini-2.5-flash

Balanced performance

2.5 Flash-Lite

gemini-2.5-flash-lite

High-throughput + low cost

2.5 Flash Live

Preview ID

Streaming interactions

2.5 TTS Variants

Preview

Speech synthesis

2.5 Audio Dialog

Preview

Native audio workflows

··········

··········

AI Studio includes the full Gemini 2.0 series and Flash/Flash-Lite models for backward compatibility and cost-optimized deployments.

Developers requiring stable legacy behavior or backward-compatible pipelines can access the complete Gemini 2.0 model set, including Flash and Flash-Lite variants still supported across Google Cloud environments.

These models provide cost-efficient paths for migration or long-running applications that depend on consistent output patterns, predictable latency profiles and stable multimodal behavior across earlier generations.

Gemini 2.0 Flash supports multimodal reasoning across text, code and images, while 2.0 Flash-Lite offers even lower inference cost for lightweight workloads and high-frequency operations.

AI Studio maintains these models to ensure compatibility with existing applications and to support gradual upgrades toward the 2.5 generation when desired.

·····

Gemini 2.0 and Legacy Models

Model Name

Identifier

Primary Use Case

2.0 Flash

gemini-2.0-flash

Multimodal baseline

2.0 Flash-Lite

gemini-2.0-flash-lite

Low-latency workloads

Other 2.0 Variants

Legacy IDs

Existing pipelines

··········

··········

Google AI Studio exposes multimodal generators for images, video and audio that extend the Gemini language models with creative and production workflows.

Alongside language models, AI Studio includes access to Google’s image-generation models such as Imagen, supporting high-fidelity image synthesis, multi-step visual transformations and structured visual control prompts.

The platform also provides access to Veo and other video-generation systems in supported regions, enabling developers to prototype short clips, animated sequences or video-based creative outputs within Studio’s experiment interface.

Audio and speech-related models, including TTS and audio-dialogue variants within the Gemini family, enable developers to build multimodal agents, real-time dialogues, voice-driven applications and speech enhancement workflows integrated with Google Cloud’s media stack.

This multimodal coverage allows developers to build unified applications that combine reasoning, vision, audio and generation under a single API ecosystem.

·····

Creative & Multimodal Model Availability

Model Type

Examples Available in AI Studio

Use Case

Image Models

Imagen variants

Image synthesis

Video Models

Veo (where available)

Video generation

Audio/TTS Models

TTS Preview

Speech output

Audio Dialog Models

Native audio dialog

Voice assistants

Multimodal Fusion

Gemini 2.5 family

Vision + text reasoning

··········

··········

Open-model support through Gemma provides lighter-weight, research-friendly models for experimentation and local or hybrid deployment.

Google includes the Gemma model family in AI Studio, designed for research environments, edge-device experimentation, fine-tuning tasks and hybrid-cloud deployments.

Gemma models provide reduced parameter footprints suitable for classroom use, prototype systems, constrained environments or developers requiring substantial flexibility beyond large-model inference.

Gemma models can be used in AI Studio for experimentation and then exported or integrated into local systems or hybrid deployments using frameworks compatible with open-model architectures.

This dual support for proprietary Gemini models and open Gemma variants widens the platform’s accessibility across domains and technical requirements.

·····

Gemma Model Access

Gemma Variant

Profile

Developer Benefit

Gemma 2 / 2B / 7B

Lightweight

Research and prototyping

Gemma Code Variants

Code-optimized

Development workflows

Fine-Tuned Gemma

Task-specific

Local deployment

··········

··········

The Gemini API provides a programmatic model listing endpoint, enabling real-time discovery and filtering of all models available through AI Studio.

Developers can retrieve Google’s full model catalog using the models.list endpoint, which returns metadata including model IDs, modality support, context windows, generation limits, preview or stable status and region availability.

This dynamic listing mechanism ensures production systems can check model availability, migrate workloads to newer variants or ensure compatibility across distributed deployments.

AI Studio mirrors this programmatic listing inside its interface, allowing developers to switch models directly in the experiment workspace and inspect documentation linked to each model type.

The unified listing system supports consistent model identification across AI Studio, Vertex AI, Cloud Console and direct API integrations.

·····

Model Listing Metadata

Metadata Field

Description

Developer Use

Model ID

Unique identifier

API calls

Modality Support

Text, vision, audio

Capability check

Context Window

Max tokens accepted

Prompt design

Output Limits

Generation ceilings

Response planning

Model Status

Stable / Preview

Deployment decisions

··········

··········

Model selection in AI Studio depends on reasoning depth, latency tolerance, multimodal needs and cost considerations.

Developers can choose between model tiers by evaluating reasoning complexity, expected input size, response-time sensitivity, operational budget and pipeline design.

Pro models suit analytical, research-heavy, deep reasoning workflows that depend on long-context depth and complex multimodal chains.

Flash models balance accuracy with cost and latency, forming an ideal mid-tier for most applications that require both speed and capability.

Flash-Lite models provide the lowest inference cost and highest throughput, supporting microservice architectures, mobile applications, event-driven systems and agentic workflows.

Open models like Gemma serve experimental, edge-device and customizable deployments where developers require full control over inference environments.

·····

Model Selection Framework

Use Case Category

Best Model Tier

Rationale

Deep Reasoning

Gemini 2.5 Pro

Highest capabilities

Balanced Workloads

Gemini 2.5 Flash

Strong multimodality

High-Throughput

Gemini 2.5 Flash-Lite

Cost-focused

Real-Time Dialog

Flash Live / Audio Dialog

Streaming behavior

On-Device / Research

Gemma

Open deployment

··········

FOLLOW US FOR MORE

··········

··········

DATA STUDIOS

··········

bottom of page