Vercel Sandbox now supports creating sandboxes with up to 32 vCPUs and 64 GB of RAM for Enterprise customers. This enables running large, resource-intensive applications that are CPU-bound or require a large amount of memory.
Get started by setting the resources.vcpus option in the SDK:
Fast mode support for Claude Opus 4.6 is now available on AI Gateway.
Fast mode is a premium high-speed option that delivers 2.5x faster output token speeds with the same model intelligence. This is an early, experimental feature.
Fast mode's increased output token speeds enable new use cases, especially for human-in-the-loop workflows. Run large coding tasks without needing to context switch and get planning results without extended waits.
To enable fast mode, pass speed: 'fast' in the anthropic provider options in AI SDK:
import{ streamText }from"ai";
const{ text }=awaitstreamText({
model:'claude-opus-4-6',
prompt:
`Analyze this codebase structure and create a step-by-step plan
to add user authentication.`,
providerOptions:{
anthropic:{
speed:'fast',
},
},
});
You can use fast mode with Claude Code via AI Gateway by setting "fastMode": true in your settings.json.
{
"model":"opus[1m]",
"fastMode":true
}
Try fast mode directly in the AI Gateway playground for Opus 4.6.
Fast mode is priced at 6x standard Opus rates.
Standard
Fast Mode
Input: $5 / 1M tokens Output: $25 / 1M tokens
Input: $30 / 1M tokens Output: $150 / 1M tokens
All standard pricing multipliers (e.g., prompt caching) apply on top of these rates.
Designed for long-horizon autonomous tasks, GLM-5.1 can work continuously on a single task for extended periods, handling planning, execution, testing, and iterative refinement in a closed loop. Rather than one-shot code generation, it runs an autonomous cycle of benchmarking, identifying bottlenecks, and optimizing across many iterations, with particular strength in sustained multi-step engineering workflows.
Beyond agentic coding, GLM-5.1 improves on general conversation, creative writing, front-end prototyping, and office productivity tasks like generating PowerPoint, Word, and Excel documents.
To use GLM 5.1, set model to zai/glm-5.1 in the AI SDK.
import{ streamText }from'ai';
const result =streamText({
model:'zai/glm-5.1',
prompt:
`Refactor the data ingestion pipeline to support streaming,
add error recovery, and benchmark throughput against the
current implementation.`,
});
AI Gateway provides a unified API for calling models, tracking usage and cost, and configuring retries, failover, and performance optimizations for higher-than-provider uptime. It includes built-in custom reporting, observability, Bring Your Own Key support, and intelligent provider routing with automatic retries.
Observability Plus's query builder now lets you create custom queries on workflow runs and steps, visualizing traffic, performance, and other key metrics across Vercel Workflows.
Queries include breakdowns by run and step status, and can be filtered and grouped by environment, project, workflow, and step.
The query builder is available to Pro and Enterprise teams using Observability Plus.
Vercel Microfrontends now include two new setup and management tools: an AI skill for coding agents and new Vercel CLI commands.
New Vercel Microfrontends skill: Install the Microfrontends skill to let your AI coding agent guide you through group creation with natural language prompts. It will automatically generate microfrontends.json, wire up framework integrations, and manage projects, all without leaving your editor.
npx skills add vercel/microfrontends
Once added, ask your agent to create your first microfrontend group using this prompt.
New CLI commands: The Vercel CLI now includes commands for managing microfrontend groups, so you can create, inspect, and manage groups from the terminal without opening the dashboard.
AI Gateway now supports Zero Data Retention (ZDR) at the team level, removing the need to configure opt-outs or reach agreements with each provider individually. It routes requests only to providers where ZDR agreements are in place, with support for Anthropic, OpenAI, Google, and many more models.
Enable team-wide ZDR from the AI Gateway Dashboard Settings to enforce Zero Data Retention on every request your team makes, with no code changes required.
You can also configure ZDR and/or No Prompt Training on the request level by setting zeroDataRetention: true or disallowPromptTraining: true per request. Enabling Zero Data Retention automatically includes training opt-out, so there's no need to configure both separately.
Commands like vc project ls and vc domains ls now automatically use the scope of your linked local directory instead of defaulting to your global team.
Previously, querying projects or domains inside a linked repository would return global results, creating an unexpected disconnect between your immediate working environment and the CLI output. This update aligns read-only commands with your local context, though you can still manually override the target team by passing the --scope flag.
Run pnpm i -g vercel@latest to update to the latest Vercel CLI (at least v50.40.0).
You can now run SQL queries, view and edit data, and inspect your database schema directly from the Vercel dashboard. This capability is available for supported marketplace database integrations, including AWS Aurora Postgres, Neon, Prisma, and Supabase, with more services coming soon.
Instead of relying on psql or external database UI tools, teams can manage their data directly within Vercel.
The integration resource page for supported Postgres databases now includes three new tabs:
Query: Run any SQL query, view results, and copy them as CSV, JSON, or Markdown.
Data Editor: View table data in a spreadsheet-like interface. You can sort, copy, edit, insert, and delete rows, then apply changes to the database atomically.
Schema: View tables and relations in a visual graph layout.
This is currently available to users with Owner permissions. Manage your database by navigating to the Browser section of your database page in the dashboard.