NVIDIA Intros Personal AI Supercomputers

NVIDIA has introduced a new lineup of AI-powered computing solutions designed to accelerate enterprise workloads. The NVIDIA DGX Spark and NVIDIA DGX Station personal AI supercomputers, powered by the NVIDIA Grace Blackwell platform, provide AI developers, researchers and data scientists with powerful tools to prototype, fine-tune and deploy large-scale AI models from their desktops.

"AI has transformed every layer of the computing stack. It stands to reason a new class of computers would emerge — designed for AI-native developers and to run AI-native applications," said Jensen Huang, founder and CEO of NVIDIA. "With these new DGX personal AI computers, AI can span from cloud services to desktop and edge applications."

DGX Spark, previously known as Project DIGITS, is billed as the world's smallest AI supercomputer, featuring the NVIDIA GB10 Grace Blackwell Superchip. It integrates a Blackwell GPU with fifth-generation Tensor Cores and FP4 precision, offering up to 1,000 trillion operations per second of AI compute power. This enables researchers to push the boundaries of generative AI, robotics, and scientific computing, according to the company.

Meanwhile, the DGX Station is the first desktop system built with the NVIDIA GB300 Grace Blackwell Ultra Desktop Superchip, delivering 784 GB of coherent memory. It is optimized for enterprise AI development, featuring the NVIDIA ConnectX-8 SuperNIC, which enables 800 Gb/s networking speeds for large-scale, multi-system AI workloads.

Both systems are designed for migration between local computing and cloud-based AI workloads, allowing users to develop and scale applications across NVIDIA DGX Cloud and other datacenter infrastructures. Manufacturing partners such as ASUS, Dell Technologies, HP and Lenovo will offer the new systems later this year.

NVIDIA Announces Blackwell RTX PRO

In addition to its personal AI computing solutions, NVIDIA introduced the RTX PRO 6000 Blackwell Server Edition, the first Blackwell-powered datacenter GPU designed for both AI and graphics-intensive enterprise workloads. With 96GB of GDDR7 memory and support for Multi-Instance GPU (MIG) technology, the RTX PRO 6000 Blackwell Server Edition will enable enterprises to securely partition workloads, improving efficiency for AI and graphics processing.

"Based on early results, we expect great performance from the RTX PRO 6000 Blackwell Server Edition," said Kris Bhaskar, senior fellow and vice president of AI initiatives at KLA. "The increased memory capacity, FP4 reduced precision and new computational capabilities of NVIDIA Blackwell are going to be particularly helpful to KLA and its customers."

RTX PRO Blackwell Series for Workstations and Laptops

NVIDIA also unveiled a lineup of RTX PRO Blackwell GPUs for desktops and mobile workstations. These GPUs cater to professionals in AI, content creation, engineering and design, offering ray tracing, neural rendering, and AI inferencing capabilities.

The new workstation lineup includes:

  • RTX PRO 6000 Blackwell Workstation Edition (24GB VRAM)
  • RTX PRO 5000, 4500, and 4000 Blackwell GPUs
  • RTX PRO 6000 Blackwell Max-Q for laptops
  • RTX PRO 5000, 4000, 3000, 2000, 1000, and 500 Blackwell GPUs for mobile workstations

These GPUs introduce new Streaming Multiprocessors with fourth-generation RT Cores and fifth-generation Tensor Cores supporting FP4 precision and DLSS 4 Multi Frame Generation.

For more information, visit the NVIDIA site.

About the Author

Chris Paoli (@ChrisPaoli5) is the associate editor for Converge360.

Featured

  • A person types on a laptop surrounded by floating digital profiles

    CrowdStrike to Acquire Identity Security Firm SGNL

    AI-based cybersecurity firm CrowdStrike has signed a definitive agreement to acquire identity security company SGNL in a deal valued at approximately $740 million.

  • Interface buttons of Generative AI tool

    Report: No Foolproof Method Exists for Detecting AI-Generated Media

    Microsoft has released a new research report warning that no single technology can reliably distinguish AI-generated content from authentic media, and that deepening reliance on any one method risks misleading the public.

  • AI word on microchip and colorful light spread

    Microsoft Unveils Maia 200 Inference Chip to Cut AI Serving Costs

    Microsoft recently introduced Maia 200, a custom-built accelerator aimed at lowering the cost of running artificial intelligence workloads at cloud scale, as major providers look to curb soaring inference expenses and lessen dependence on Nvidia graphics processors.

  • Abstract speed motion blur in vibrant colors

    3 Ed Tech Shifts that Will Define 2026

    The digital learning landscape is entering a new phase defined by rapid advances in artificial intelligence, rising expectations for the student experience, and increasing pressure to demonstrate quality and accountability in online education.