Meta AI: Rollout updates for advanced models and feature expansions
- Graziano Stefanelli
- Aug 20
- 3 min read

Meta AI has accelerated its deployment of higher-tier Llama models, introducing larger context windows, faster throughput, and expanded multimodal capabilities. The rollout strategy is structured to move from tightly controlled design-partner programs to general availability, ensuring stability and compliance while delivering performance improvements to all user tiers.
The new model rollout follows a phased approach.
This measured sequence reduces disruption and provides Meta with actionable telemetry from early-stage users before large-scale release.
New models bring higher limits and faster execution.
These upgrades allow more complex projects to be handled in a single conversation, reducing the need for manual summarisation or splitting inputs into multiple sessions.
Availability varies by plan and location.
Deep Think remains opt-in for most plans, with Ultra restricted to a small number of enterprise partners.
Governance tools accompany each rollout.
These safeguards enable organisations to adopt new models without losing oversight of usage or compliance.
Performance benchmarks show tangible improvements.
The Deep Think model trades slightly slower first-token latency for longer memory and reasoning depth, making it more effective for extended projects.
Known issues are tracked with workarounds.
Meta publishes these limitations with recommended mitigation steps in the model release notes.
The roadmap points toward even larger context and customisation.
Planned upgrades include a fine-tuning toolkit for Llama 4 Turbo, enabling LoRA adaptation for up to 25 million tokens; on-device inference for Ray-Ban smart glasses using quantised Llama 3.5; and a code execution sandbox inside chats for Deep Think. These features are expected to extend both the autonomy and portability of the models in production environments.
____________
FOLLOW US FOR MORE.
DATA STUDIOS




