NVIDIA Secures Early Buyers for Vera CPU Racks

NVIDIA announced the Vera CPU, a purpose-built Arm-based data center processor aimed at agentic AI workloads, and has already signed multiple early deployment partners. According to NVIDIA's press release and reporting by DatacenterD and Wccftech, early collaborators and buyers include CoreWeave, Meta, Oracle Cloud Infrastructure, Alibaba Cloud, ByteDance, Lambda, Nebius, and Nscale. NVIDIA's materials and DatacenterD report the chip contains 88 custom Olympus cores, uses LPDDR5X memory, delivers up to 1.2 Tbps of bandwidth per chip, and the company announced a 256-processor liquid-cooled Vera rack that NVIDIA states can sustain more than 22,500 concurrent CPU environments. NVIDIA's press release also claims Vera delivers twice the efficiency and 50% faster performance than traditional rack-scale CPUs. Multiple OEMs and system vendors are listed as manufacturing partners in NVIDIA's release and industry reporting.
What happened
According to NVIDIA's press release and contemporaneous reporting by DatacenterD, Wccftech and StorageReview, NVIDIA launched the Vera CPU, a custom Arm-based data center processor positioned for agentic AI and reinforcement-learning workloads. NVIDIA's press release lists early collaborators and customers as CoreWeave, Meta, Oracle Cloud Infrastructure, Alibaba Cloud, ByteDance, Lambda, Nebius and Nscale, and DatacenterD and Wccftech report additional deployments at national laboratories and cloud providers. NVIDIA announced a rack configuration integrating 256 liquid-cooled Vera CPUs that, per NVIDIA's materials, can sustain more than 22,500 concurrent CPU environments.
Technical details
DatacenterD reports the Vera design contains 88 custom Olympus cores with NVIDIA Spatial Multithreading, uses LPDDR5X memory, and delivers up to 1.2 Tbps of memory bandwidth. NVIDIA's press release states the architecture targets higher single-thread performance and claims Vera provides twice the energy efficiency and 50% faster performance than traditional rack-scale CPUs. System makers named in NVIDIA's announcement and industry coverage include Dell Technologies, HPE, Lenovo, Supermicro, and a broad set of OEM partners identified in NVIDIA's materials.
Industry context
Editorial analysis: The coverage places Vera inside a broader industry shift toward CPU-attentive infrastructure for low-latency, multi-tasking agentic workloads. Public reporting from GTC 2026 and follow-up pieces frame Vera as part of NVIDIA's rack-scale strategy alongside Rubin GPUs and Groq LPUs, reflecting a trend where vendors deliver integrated chip-to-rack systems rather than discrete components.
Context and significance
Editorial analysis: Multiple cloud providers and hyperscalers appearing on the early-adopter lists is notable because it signals demand for CPU resources that coordinate models, run toolchains, and host many concurrent agentic environments. Reporting by DatacenterD and other outlets highlights both standalone Vera offerings and Vera paired with NVIDIA GPUs, which could broaden choices for operators balancing inference latency, throughput, and energy cost.
What to watch
Editorial analysis: Observers should track benchmark and independent performance data for Vera-based racks versus incumbent x86 and Arm offerings, actual availability timelines from named cloud providers, and which OEMs ship validated Vera systems. Also monitor how Vera adoption affects rack-level power, cooling, and orchestration patterns in production AI factories. For practitioners: measure latency, context-switch behavior under multi-tenant agent workloads, and memory bandwidth utilization in pilot deployments before large-scale migration decisions.
"Vera is arriving at a turning point for AI. As intelligence becomes agentic, capable of reasoning and acting, the importance of the systems orchestrating that work is elevated," Jensen Huang said in NVIDIA's press release. DatacenterD and StorageReview provide complementary technical summaries and deployment lists cited above.
Scoring Rationale
This is a notable infrastructure release because multiple hyperscalers and cloud providers are named as early buyers, and Vera is presented as a rack-scale building block for agentic AI. The story matters to practitioners planning large-scale inference or agent deployments but is not a frontier-model breakthrough.
Practice interview problems based on real data
1,500+ SQL & Python problems across 15 industry datasets — the exact type of data you work with.
Try 250 free problems
