Arrcus just announced a tight integration between its Arrcus Inference Network Fabric (AINF) and NVIDIA’s AI stack at NVIDIA GTC including the NVIDIA Dynamo framework, BlueField-3 DPUs, Spectrum-X Ethernet, and GPU platforms.

Powered by NVIDIA tech, AINF creates a secure, smart, policy-driven inference fabric that stretches from edge to data center to cloud. Companies like Lightstorm can now roll out next-gen Physical and Agentic AI apps with lower latency, better GPU efficiency, less power draw, and cheaper per-inference costs.
AI’s massive shift from centralized training to distributed global inference is shaking up infrastructure fast. Agentic workflows chain together dozens of model calls and tools needing instant model picking, priority handling, and policy checks right at the network edge.
Real-time apps like robotics, autonomy, video analytics, they crave ultra-low latency, geo-smart routing, clever model selection, GPU optimization, and secure multi-site links.
Old-school networking and load balancers just can’t hack it at global scale. Arrcus + NVIDIA fixes that with a seamless inferencing fabric, where AINF plays central conductor using smart LLM classifiers on NVIDIA gear to pick the best model and route requests to the ideal site/path in real time.
This unlocks:
- Lightweight models at the edge for quick tasks
- Centralized heavy reasoning where it shines
- Priority lanes for voice/video/gaming
- Network slicing for rock-solid SLAs
Arrcus AINF + NVIDIA Dynamo
AINF pairs with Dynamo as the global routing brain for inference—steering requests based on model availability, service tiers, geofencing, capacity, and network health. Dynamo handles local GPU load balancing per site, while AINF owns cross-site decisions.
- Dynamo feeds GPU stats (queue depth, KV-cache, replica health)
- AINF’s Site Agent pulls it into the control plane
- AINF crunches global, model-aware routes
Perfect combo: optimal site + best replica = max compute bang.
AINF + BlueField & Spectrum
AINF taps NVIDIA BlueField-3 DPUs for locked-down inference traffic line-rate 400 Gb/s encryption with zero CPU hit, keeping sovereign data safe across sites. Paired with Spectrum-4 Ethernet switches and GPUs, it’s an end-to-end fabric built for speed, efficiency, and security.
Catch the live demo of AINF + NVIDIA AI-Grid at NVIDIA GTC booth #3019 (March 16–19). Book time with Arrcus at their GTC meeting page. Game-changer for distributed AI inference!
Leadership Comments
“AI is entering its inference era, where networking becomes the control plane for performance and economics,” said Shekar Ayyar, Chairman and CEO of Arrcus. “By integrating AINF with NVIDIA AI technologies, we are enabling operators and enterprises to intelligently route inference traffic, maximize GPU utilization and deliver real-time AI services at global scale.”
“AI inferencing at scale across Asia?Pacific demands reliable, low?latency connectivity across vast WAN distances. Lightstorm is enabling hyperscalers, neoclouds and enterprises with the network foundation required for this shift, and by leveraging Arrcus’ AINF solution powered by NVIDIA, we’re excited to make real?time, large?scale inferencing into a deployable reality in the region,” said Amajit Gupta, Group CEO & MD, Lightstorm.





