Grok: rollout updates and feature roadmap for advanced models
- Graziano Stefanelli
- Aug 28
- 3 min read

Grok’s ecosystem has evolved rapidly, with a growing catalog of models, new licensing terms, and updated performance limits across the free and paid tiers. The rollout includes open-weight releases, new context capabilities, roadmap timelines, and integration improvements for both developers and enterprise teams.
Grok-1 open-weight release expands adoption possibilities.
In March, Grok-1 became one of the first large-scale models from xAI released under the Apache-2.0 open-source license. The initial release on GitHub included checkpoints, model cards, and tokenizer configurations, enabling direct deployment on private infrastructure without vendor lock-in.
The original tokenizer supported 16 000 tokens, but an update in August expanded the context window to 24 000 tokens, giving developers more flexibility when working with extended datasets, multi-document analysis, and conversation threads that require higher retention.
Nano-series models bring efficiency for lighter workloads.
The Nano series was introduced to cover lower-latency, energy-efficient use cases. Currently, the available models include:
Model | Parameters | Context Window | Release Date | License | Notable Features |
Nano-7B | 7B | 16 000 tokens | 4 Jun 2025 | Apache-2.0 | Optimized for mobile and edge workloads |
Nano-20B | 20B | 16 000 tokens | 4 Jun 2025 | Apache-2.0 | Best performance-per-token ratio |
Nano-40B (roadmap) | 40B | 32 000 tokens | Q4 2025 | Apache-2.0 | Designed for enhanced multimodal support |
The release also introduced quantized checkpoints for lower memory requirements, improving accessibility for developers working on embedded devices.
Grok µ-2B offers high-speed, low-cost experimentation.
Released in July, Grok µ-2B delivers fast response times with 2 billion parameters while maintaining accuracy on basic tasks. It's designed for research workflows, rapid prototyping, and smaller-scale assistants.
The model’s compact size significantly reduces compute demands, making it easier to integrate into constrained environments where latency and cost are critical factors.
Grok 4 upgrades context windows and pricing options.
The release of Grok 4 marked a major milestone for xAI’s advanced model strategy. The model supports a 128 000-token context window, enabling extended reasoning over larger documents, codebases, and multi-turn conversations.
Initially, the free tier offered 3 000 000 tokens per month, but as of 7 August, the quota has been reduced to 2 000 000 tokens per month. Paid plans provide additional throughput and higher token caps for enterprise-scale workflows.
Tier | Monthly Token Limit | Throughput (TPS) | Cost | Availability |
Free | 2 000 000 | 10 | $0 | Web + API |
Pro | 8 000 000 | 25 | $20/month | Web + API |
Enterprise | Custom | 50+ | Negotiated | API only |
This tiered system allows individual users, startups, and large enterprises to adopt Grok at different scales depending on performance and budget needs.
Vision expansion introduces multimodal capabilities.
The rollout also introduced Grok-V Mini, a 9B-parameter vision adaptor supporting multi-image input with a 32 000-token context. Released under a CC-BY-NC license, this version is optimized for image recognition, chart parsing, and document classification tasks.
Upcoming enhancements scheduled for November include support for simultaneous multi-frame video inputs and improved integration with PDF analysis pipelines.
Roadmap points to Nano-40B and Grok 4 Heavy.
The near-term development plan includes two notable launches:
Nano-40B – Scheduled for Q4 2025, offering 32 000 tokens of context and better multimodal performance.
Grok 4 Heavy – A premium variant entering trial access in October 2025, designed for high-throughput enterprise orchestration and enhanced retrieval-augmented generation workflows.
These developments position Grok to compete more directly with other advanced AI models, expanding both performance headroom and deployment flexibility.
Integration improvements with Azure and enterprise platforms.
For a limited period between June and July, Grok 4 was bundled with Azure AI Foundry, allowing developers to provision GPU-backed environments at discounted rates. Although the promotion has ended, API integration remains fully supported via Azure and independent endpoints.
Enterprise adoption has grown due to Grok’s open-weight strategy, giving organizations the ability to self-host models or leverage xAI’s infrastructure for managed deployments.
This release cycle marks one of Grok’s most comprehensive updates, combining open-source accessibility with advanced multimodal performance. With growing free-tier availability, better pricing flexibility, and new features on the roadmap, the Grok ecosystem is positioned for broader developer and enterprise integration.
____________
FOLLOW US FOR MORE.
DATA STUDIOS