CIO Influence
CIO Influence News Machine Learning Networking

Arrcus Delivers Record Breaking 3x Bookings Growth in 2025, and Introduces AI-Policy Aware Arrcus Inference Network Fabric

Arrcus Delivers Record Breaking 3x Bookings Growth in 2025, and Introduces AI-Policy Aware Arrcus Inference Network Fabric

Arrcus Logo

Purpose-built network fabric designed to accelerate delivery of real-time and agentic AI applications with improved throughput and power efficiency while reducing token retrieval time, latency, and cost per inference.

Arrcus, the leader in distributed networking infrastructure, announced record 3x bookings growth in 2025 across datacenter, telco and enterprise customers for mission critical switching and routing applications deployed in production across thousands of network nodes globally. Customers have valued the flexibility, innovation and feature velocity of the ArcOS® network operating system and the ACE® platform across a broad range of open networking hardware, along with significant reduction in capital and operating costs compared to alternative incumbent networking solutions. Building on this success, the company also announced its Arrcus Inference Network Fabric (AINF), designed to improve the delivery of inferencing AI applications across a highly distributed network by steering traffic between inferencing nodes, caches, and datacenters with the goal of increasing throughput Tokens per second (TPS), reducing Time to First Token (TTFT), and improving End to End Latency (E2EL) for inferencing.

Also Read: CIO Influence Interview With Jake Mosey, Chief Product Officer at Recast

“AINF extends Arrcus’ leadership in distributed networking by delivering the first fabric designed to meet the latency, sovereignty, and power constraints of large-scale AI inferencing,” said Shekar Ayyar, Chairman and CEO of Arrcus.

With the rise of Agentic and Physical AI, Inferencing is expected to be the fastest growing AI segment. However, widespread adoption of Agentic AI is bottlenecked by challenges in the speed of delivery of inference results, diversity in inference models, and bringing smart inference decision making closer to edge nodes. Inferencing infrastructure is deployed in highly distributed clusters, and needs to address the requirements of low latency, availability, constraints in power grid capacity, data sovereignty, and cost. While Enterprises are looking to deploy real-time inferencing so users can have rich localized experiences, Network operators are looking to deliver Inferencing-as-a-service in alignment with the Service Level Objectives (SLO) around these key requirements. To meet these challenges, inferencing infrastructure will require a distributed routing fabric that has granular policy control to intelligently steer traffic and match rapidly evolving requirements. Traditional hardware-defined networking solutions from incumbent vendors fall short in addressing these challenges.

Announced , the Arrcus Inferencing Network Fabric (AINF) is a purpose-built solution that enables delivery of Inferencing applications with an intelligent ‘AI policy-aware’ network fabric that can dynamically route AI traffic as required between inference nodes, caches, and datacenters to the most appropriate site. Operators can define business policies such as latency targets, data sovereignty boundaries, model preferences, or power constraints. AINF enables evaluation of these conditions in real time to steer inference traffic to the optimal node or cache, ensuring the right model is delivered from the right location at the right time. Research1 shows that such innovation in AI infrastructure can deliver over 60% reduction in TTFT, 15% TPS improvement, 40% E2EL and up to 30% cost reduction.

“To enhance agentic AI adoption by improving response times, networks need to become AI-aware,” said Shekar Ayyar, Chairman and CEO of Arrcus. “AINF extends Arrcus’ leadership in distributed networking by delivering the first fabric designed to meet the latency, sovereignty, and power constraints of large-scale AI inferencing.”

At its core, AINF introduces a policy abstraction layer that translates the inferencing application intent to underlying infrastructure performance, while shielding operators from infrastructure complexity. AINF components include query-based inference routing with policy management, interconnect routers, and edge networking. AINF is designed to integrate with popular inference frameworks including vLLM, SGLang, Triton, and others, thus coupling optimal model selection with a high performance steering fabric. Using Kubernetes-based orchestration AINF can be composed and deployed in an automated manner. Concepts like prefix awareness to optimize KV cache usage enable inferencing applications to meet SLO for throughput, token retrieval time, latency, data sovereignty, power and cost.

AINF builds on Arrcus’ proven leadership in AI and Datacenter Networking, with its ACE-AI solution already delivering a unified network fabric for distributed AI spanning datacenters, edge and hybrid cloud environments with scale-out and scale-across solutions. As with all Arrcus solutions, AINF has the unique capability of working with best of breed inferencing xPUs and network silicon across hardware providers. It is also designed to allow partner companies to bring in their load balancers, firewalls, and power management policies to create optimal caching and secure CDNs for superior inference results.

Catch more CIO Insights: Why CIOs are becoming chief risk orchestrators?

[To share your insights with us, please write to psen@itechseries.com ]

Related posts

Fungible Announces Fungible Storage Cluster 4.0 Empowering Tomorrow’s Data Center

Reply Launches Silicon Shoring: an AI-Powered Software Delivery Model to Optimise and Automate the Entire Software Development Life Cycle

Business Wire

Zilliz Cloud Delivers Major Cost Savings, Higher Performance, and Strengthened Security for Enterprise AI

PR Newswire