Editorial illustration for NVIDIA DGX Spark expands node support to four, doubling memory capacity
NVIDIA DGX Spark Expands Memory for AI Workloads
NVIDIA DGX Spark expands node support to four, doubling memory capacity
Why does the memory ceiling matter for autonomous AI agents? Those models chew through data fast, and a single DGX Spark board caps out at 128 GB of RAM. When researchers tried to push larger workloads, they hit a wall: the platform could only double that headroom by linking two nodes, reaching 256 GB.
That limit forced teams to split jobs or compromise on model size, slowing experiments that already run at the edge of feasibility. Here's the thing: expanding the cluster beyond two units changes the calculus entirely. More nodes mean not just more memory, but new ways to arrange compute—different execution topologies that can be matched to specific tasks.
The upgrade promises a broader canvas for scaling agents that need to reason across massive state spaces, and it could reshape how labs benchmark multi‑node performance. In short, the next step opens the door to workloads that were previously out of reach, setting the stage for the announcement that follows.
Until now, NVIDIA DGX Spark has supported scaling up to two nodes, increasing the available memory from 128 GB on one node to 256 GB on two nodes. This capability has now been increased to up to four DGX Spark nodes. DGX Spark also now supports several execution topologies, each tailored to different goals through the low latency of RoCE communication enabled by ConnectX-7 NICs.
- One DGX Spark node: Ideal for low latency, large context size inference, fine-tuning up to 120B parameters, and local agentic workloads - Two DGX Spark nodes: Balanced scaling for faster fine-tuning and larger models, as well as support for up to 400B-parameter inference - Three DGX Spark nodes in a ring: Ideal for fine-tuning larger models or small training jobs - Four DGX Spark nodes with RoCE 200 GbE switch: Local inference server ideal for state-of-the-art models up to 700B parameters, communication intensive workloads, and local AI factory operations Inference can scale up linearly on DGX Spark when internode communication is minimal. When work is largely independent per GPU, the results are aggregated once at the end rather than continuously. In this case, DGX Spark nodes can run in parallel with low synchronization overhead.
For example, a reinforcement learning (RL) workload in NVIDIA Isaac Lab can run many simulations independently on each node. Results are collected in a single step, yielding near-linear scaling across multiple DGX Spark nodes. Inference scaling is less than linear when the workload requires frequent, fine-grained communication between nodes.
During LLM inference, model execution occurs layer by layer, with continuous synchronization required across nodes. Partial results from different DGX Spark nodes must be exchanged and merged repeatedly, which introduces significant communication overhead.
The upgrade pushes DG X Spark from a two‑node ceiling to four, effectively doubling the pool of on‑board memory that autonomous agents can draw on. Where a single node offered 128 GB, two nodes already delivered 256 GB; four nodes now promise up to 512 GB of local capacity. NVIDIA NemoClaw, part of the NVIDIA Agent Tool, is bundled with the platform, suggesting tighter integration for complex workflows.
Several execution topologies are now supported, each tailored to different workload patterns, though the specifics of those topologies remain undocumented. This expansion could ease the pressure autonomous agents face when juggling long‑running tasks, multiple communication channels, and background subprocesses. Yet, whether the added nodes will translate into proportionally faster or more efficient agent performance is unclear.
The hardware boost is tangible, but the real‑world impact on AI agent productivity will depend on software orchestration and workload characteristics. As the system scales, monitoring how memory and compute resources are allocated will be essential to validate the promised benefits.
Further Reading
- Nvidia DGX Spark Now Supports Four-Node Linking and 512GB Shared RAM - Guru3D
- NVIDIA GTC 2026: Live Updates on What's Next in AI - NVIDIA Blogs
- NVIDIA DGX Spark: AI Supercomputer on Your Desk - NVIDIA
- Nvidia GTC San Jose 2026 - Level1Techs Forums
Common Questions Answered
How has NVIDIA DGX Spark expanded its node support capabilities?
NVIDIA DGX Spark has increased its node support from two to four nodes, effectively doubling the available memory from 256 GB to 512 GB. This expansion allows researchers to run larger AI workloads and more complex autonomous agent models without previous memory constraints.
What communication technology enables the new DGX Spark node configurations?
The new DGX Spark configurations leverage RoCE (RDMA over Converged Ethernet) communication enabled by ConnectX-7 NICs, which provides low-latency connections between nodes. This technology allows for multiple execution topologies that can be tailored to different computational goals and workload patterns.
What tool is bundled with the updated NVIDIA DGX Spark platform?
NVIDIA NemoClaw, part of the NVIDIA Agent Tool, is now bundled with the DGX Spark platform, suggesting enhanced integration for complex AI workflows. This tool is likely designed to help manage and optimize the expanded multi-node computational capabilities.