/* Premium Sticky Anchor - Add to the section of your site. The Anchor ad might expand to a 300x250 size on mobile devices to increase the CPM. */ Grok AI: All Models Available: capabilities, context windows, pricing, and when to use each
top of page

Grok AI: All Models Available: capabilities, context windows, pricing, and when to use each

xAI’s Grok stack now spans frontier reasoning, cost-efficient “fast” variants, code-focused models, and legacy VLM/image engines. For builders and researchers, the practical question is no longer “Does Grok do X?” but which Grok to deploy for a given job (reasoning, coding, retrieval, or vision) and at what price point. Below is a consolidated, source-backed map of every generally available Grok model family today—plus historical models you’ll still encounter in docs and repos.

·····

.....

The current catalog (API & product)

Model (alias)

Modality

Context window

Notable features

Throughput (typical)

Pricing (per 1M)*

grok-4-fast-reasoning

Text (reasoning)

2,000,000

Tool use, structured output, function calling

~480 rpm

$0.20 in / $0.50 out

grok-4-fast-non-reasoning

Text

2,000,000

Cost-efficient drafts; no reasoning tokens

~480 rpm

$0.20 in / $0.50 out

grok-4-0709 (aka grok-4)

Text (frontier)

256,000

Highest quality “full Grok 4”

~480 rpm

$3 in / $15 out

grok-3

Text

131,072

Strong enterprise tasks; extraction & coding

~600 rpm

$3 in / $15 out

grok-3-mini

Text

131,072

Lower cost, faster than Grok-3

~480 rpm

$0.30 in / $0.50 out

grok-code-fast-1

Text (code)

256,000

Agentic coding; tool use optimized

~480 rpm

$0.20 in / $1.50 out

grok-2-vision-1212 (us-east-1 / eu-west-1)

Text+Image in

32,768

Multimodal understanding of images

up to 600 rpm (region-dependent)

$2 in / $10 out

grok-2-image-1212

Image out

Image generation

~300 rpm

$0.07 / image

*Pricing, context, and rate limits are taken from xAI’s Models & Pricing documentation; regional variants and rps/rpm differ by endpoint.

Key clarifications from the docs: Grok-4 is a reasoning model (no separate “non-reasoning mode”); some completion parameters (e.g., stop, penalties) are not supported on reasoning models. Also, real-time knowledge requires tools—by default models do not have current-events awareness without Live Search.

·····

.....

Grok 4: the frontier tier (and why “Fast” matters)

  • Grok-4-0709 (aka grok-4) is the highest-quality baseline for complex synthesis, analysis, and instruction following. It carries the premium price point and a 256k context. For maximum quality, use this.

  • Grok-4-Fast (reasoning / non-reasoning) stretches the context to 2M tokens while radically cutting price and latency. It’s designed for cost-efficient reasoning and high-throughput automations where you need long context + tool use, not necessarily the absolute peak quality of full Grok-4.

xAI positioned Grok 4 publicly as its new flagship in July 2025 (API + consumer surfaces). Premium tiers on X and SuperGrok also advertise Grok-4 access.

·····

.....

Grok 3 family: strong generalist + a smaller “mini”

xAI introduced Grok-3 and Grok-3 mini as beta reasoning models with RL-enhanced “Think” modes; they became widely available across API and cloud hosts (including Azure AI Foundry). Grok-3 targets enterprise-style tasks—data extraction, programming, summarization—and sits at the same list price as Grok-4 (standard) but with a 131k window; mini drops cost for speed.

Even after Grok-4’s launch, grok-3 remains visible in the models table for compatibility and cost/perf tradeoffs. If you need reliable JSON extraction at moderate cost and don’t need the newest tool stack, Grok-3 is a sensible default.

·····

.....

Grok Code Fast 1: agentic coding (new)

Announced late August, grok-code-fast-1 is a “lightning-fast” reasoning model built for agentic coding: tool-augmented code generation, refactors, and environment-aware planning. It pairs a 256k window with low input cost and higher output cost (reflecting longer completions for code). For developer copilots and pipeline automation, it’s the value workhorse.

·····

.....

Vision & image models: Grok-2 VLMs remain in catalog

For multimodal pipelines, the current docs still list grok-2-vision-1212 (image understanding with 32k context) and grok-2-image-1212 (image generation at $0.07/image). These are stable, production VLMs useful when you need image inputs/outputs alongside Grok-3/4 text models.

Tip: If you need PDF page or chart analysis, route pages as images to the VLM or use Grok-4 with tools that extract page text + figures, then fuse results. Tools are billed separately (see below).

·····

.....

Tools, Live Search, and costs that show up on your bill

Grok’s agentic tool calling adds separate metering for server-side tools (Web Search, X Search, Code Execution, etc.). You’re charged per 1,000 tool calls in addition to tokens; Live Search has its own per-source pricing and is being superseded by the agentic tool-calling API later this year. If you rely on real-time grounding, budget for token + tool costs.

Tool pricing examples (from docs): Web Search and X Search $10 / 1k calls; Live Search $25 / 1k sources (deprecating by Dec 15, 2025).

·····

.....

Historical & legacy models you’ll still see referenced

Model

What it was

Why you’ll still see it

Grok-1 (open weights, MoE 314B)

Base model weights and architecture released openly in Mar 2024.

For research baselines, forks, and MoE studies.

Grok-1.5

Long-context (128k) upgrade over Grok-1 with stronger math/code.

Shows up in older benchmarks and migration notes.

Grok-1.5V

First multimodal preview (documents, diagrams, screenshots).

Early VLM milestone before Grok-2 Vision.

Grok-2 / Grok-2 mini

2024 beta generation that leapfrogged 1.5; underpinned later vision & image SKUs.

Useful for understanding the performance lineage.

xAI’s public timeline shows a steady march: Grok-1 → 1.5 → 1.5V → 2 / 2 mini → 3 / 3 mini → 4, with 4-Fast and Code Fast rounding out today’s production catalog.

·····

.....

Choosing the right Grok (quick guide)

Your need

Pick this

Why

Highest-quality general reasoning & writing

grok-4-0709

Best-available quality; robust tools; 256k context.

Mass-scale automations & long inputs

grok-4-fast-reasoning

2M context, far lower cost; reasoning + tools.

Low-latency drafts (no reasoning tokens)

grok-4-fast-non-reasoning

Cheapest Grok-4 variant for quick copy/transform.

Stable enterprise extractor

grok-3

Proven JSON/CSV extraction and coding help.

Budget generalist

grok-3-mini

Good enough for many tasks at minimal cost.

Image understanding

grok-2-vision-1212

Multimodal input with regional endpoints.

Image generation

grok-2-image-1212

Simple, inexpensive image synthesis.

Agentic coding

grok-code-fast-1

Purpose-built for code + tools at speed.

·····

.....

Availability notes (consumer vs. API)

  • Consumer surfaces (Grok web / X Premium+ / SuperGrok): Grok-4 runs the user-facing assistant; Heavy tier advertises a higher-capacity Grok-4 variant for power users. Feature availability (e.g., feed steering, search grounding) is evolving across X apps.

  • API & cloud hosts: Grok-3/Grok-3 mini shipped broadly (including Azure AI Foundry), then Grok-4 and Grok-4 Fast moved into the public docs and consoles with explicit pricing and limits. Always check regional endpoints and rpm/rps on your account.

·····

.....

Practical tips before you switch models

  1. Reasoning vs. non-reasoning: Reasoning models reject certain legacy params and meter reasoning tokens separately; plan prompts and budgets accordingly.

  2. Real-time grounding: Grok models have no built-in current events—enable Web/X search tools or pass context yourself.

  3. Long context strategy: Even with 2M tokens, retrieval + section prompts remain faster and cheaper than raw dumps. The docs support cached prompt tokens for repeat runs.

  4. Migration hygiene: xAI maintains model aliases (e.g., grok-4, -latest, -<date>). Pin exact versions for reproducibility; use -latest only if you accept silent upgrades.

·····

.....

Bottom line

Today’s Grok lineup spans premium Grok-4, ultra-long-context Grok-4-Fast (reasoning & non-reasoning), workhorse Grok-3 / 3-mini, the grok-code-fast-1 coder, and Grok-2 VLMs for image understanding/generation. Pair them with agentic tools when you need real-time grounding. If you want top quality, pick Grok-4; if you want scale and cost efficiency with giant prompts, pick Grok-4-Fast; for stable enterprise extraction, Grok-3 still earns its keep. All of it is documented—with prices, limits, and tool costs—right in xAI’s Models & Pricing page.

.....

FOLLOW US FOR MORE.

DATA STUDIOS

.....

Recent Posts

See All
bottom of page