CIO Influence
CIO Influence News Datacentre Machine Learning

MinIO Powers Arm-Based Chipsets for High-Performance AI

MinIO Powers Arm-Based Chipsets for High-Performance AI

MinIO Logo (PRNewsfoto/MinIO)

Improved chipset performance and efficiency enable crucial vector operations for AI, ML and other data-intensive applications

MinIO, the leader in high-performance storage for AI, today announced compelling new optimizations and benchmarks for Arm-based chipsets powering on its object store. These optimizations underscore the relevance of low power, computationally dense chips in key AI-related tasks and demonstrates the readiness and superiority of the Arm architecture for modern AI and data processing workloads including erasure coding, bit rot protection and encryption.

Also Read: AppDirect and Centripetal Partner to Offer Intelligence Powered Cybersecurity

MinIO’s work leveraged the latest Scalable Vector Extension Version (SVE) enhancements. SVE improves the performance and efficiency of vector operations, which are crucial for high-performance computing, artificial intelligence, machine learning and other data-intensive applications.

Specifically, MinIO was able to enhance its existing implementation of its Reed Solomon erasure coding library. The result was 2x faster throughput as compared to the previous NEON instruction set implementation, as is shown in this graph:

Just as significantly, the new code uses just a quarter of the available cores (16) to consume half the memory bandwidth. Previously it required 32 cores to achieve this same memory bandwidth consumption.

Highway Hash

MinIO was also able to achieve significant performance enhancements for the Highway Hash algorithm it uses for bit-rot detection. This is frequently run during both GET (read) as well as PUT (write) operations. SVE support for the core hashing update function delivered the following results:

The performance scales linearly as the core count goes up, and for the larger block sizes, it starts to reach the memory bandwidth limit around 50 to 52 cores.

Further, SVE offers support for lane masking via predicated execution, allowing more efficient utilization of the vector units. Extensive support for scatter and gather instructions makes it possible to efficiently access memory in a flexible manner.

“Delivering performance gains while improving power efficiency is critical to building a sustainable infrastructure for the modern, intensive data processing workloads demanded by AI,” said Eddie Ramirez, vice president of marketing and ecosystem development, Infrastructure Line of Business, Arm. “Performance efficiency is a key factor in why Arm has become pervasive across the datacenter and in the cloud, powering server SoCs and DPUs, and MinIO’s latest optimizations will further drive Arm-based AI innovation for the world’s leading data enterprises.”

Also Read: CIO Influence Interview with Simon Tusha, Founder and CTO of TecFusions

The latest generation of the NVIDIA® BlueField-3 data processing unit (DPU) offers an integrated 16-core Arm-based CPU as part of the network controller card itself. This leads to simpler server designs by interfacing the Non-Volatile Memory Express (NVMe) drives directly to the networking card and bypassing any (main server) CPUs altogether. This represents an ideal pairing for software-defined technologies like MinIO. With 400Gb/s, Ethernet BlueField-3 DPUs offload, accelerate and isolate software-defined networking, storage, security and management functions. Given the criticality of disaggregating storage and compute in modern AI architectures, the pairing is remarkably powerful.

MinIO has been at the forefront of storage innovation, and its latest benchmarks reveal the impressive performance of the Arm architecture in data-intensive tasks. They follow the foundational work undertaken to optimize and benchmark Arm-based AWS Graviton2 processors against Intel Skylake chips. The latest Arm architecture, particularly with NVIDIA’s BlueField-3 DPU, demonstrates remarkable efficiency and scalability for storage workloads. This includes advanced functions such as erasure coding, encryption, and hashing, essential for AI data processing.

“As the technology world re-architects itself around GPUs and DPUs, the importance of computationally dense, yet energy efficient compute cannot be overstated,” said Manjusha Gangadharan, Head of Sales and Partnerships at MinIO. “Our benchmarks and optimizations show that Arm is not just ready but excels in high-performance data workloads. We are delighted to continue to build our partnership with Arm to create true leadership in the AI field.”

[To share your insights with us as part of editorial or sponsored content, please write to psen@itechseries.com]

Related posts

Identiv Introduces Intuitive IoT Product Advisor

Business Wire

BigID Prepares Organizations for CPRA Compliance with an Automated Data Privacy Suite

PR Newswire

AFL Chooses BIAS to Spearhead its Digital Transformation Strategy to Oracle Cloud

CIO Influence News Desk