Site icon Cloudian

Breakthrough: Cloudian HyperStore Delivers 200GB/s Object Storage Performance for AI Workloads

Cloudian object storage performanceIn large-scale enterprise AI, storage performance has emerged as a critical factor in maximizing GPU utilization and accelerating time-to-insight. Today, we are pleased to share groundbreaking object storage performance results from our recent testing of Cloudian HyperStore in the context of our Rack-Scale AI Architecture.

Setting New Performance Standards for AI Storage

Our testing confirms that Cloudian HyperStore achieves an impressive 200 GiB/s read throughput on a modest 6-node cluster (12U total). This object storage performance translates to nearly 35 GiB/s per storage node

These results demonstrate that Cloudian HyperStore not only meets but significantly exceeds NVIDIA’s recommended storage performance guidelines for GPU-intensive AI workloads. With NVIDIA recommending 1.5 GB/s per GPU, our 6-node configuration delivers enough throughput to support over 130 GPUs—far more than the 32 GPUs in our reference architecture’s baseline configuration.

Testing Methodology

To ensure our object storage performance measurements reflect real-world workloads, we conducted comprehensive testing using the following configuration:

Test Parameters
Network Infrastructure
Download the full reference architecture here. 

This testing environment closely mirrors the Cloudian-Supermicro Full AI Stack Reference Architecture, providing a validated blueprint for organizations looking to deploy rack-scale AI infrastructure.

Scalability: The Key to AI Infrastructure Growth

With linear scalability of the Cloudian HyperStore architecture, this object storage performance grows as organizations add storage nodes to their clusters. This allows an organization to achieve proportional increases in throughput—scaling to terabytes per second in larger deployments.

This linear scaling characteristic is particularly valuable for AI workloads, which typically grow in both data volume and computational demands over time. Unlike traditional storage architectures that can only be scaled to fixed limits, Cloudian’s peer-to-peer, shared nothing architecture allows organizations to start with a right-sized deployment and expand seamlessly as their AI initiatives mature.

Implications for Enterprise AI Initiatives

For data scientists, ML engineers, and IT architects, these object storage performance results translate to tangible business benefits:

    1. Maximized GPU Utilization: Ensures expensive GPU resources are never idle waiting for data
    2. Reduced Training Time: Faster data delivery means more training iterations and quicker model convergence
  1. Cost-Efficient Scaling: Object storage economics combined with high performance eliminates the need for expensive specialized file systems
  2. Simplified Infrastructure: S3-compatible object storage with RDMA acceleration delivers file storage-level performance with cloud-native simplicity

The Technology Behind the Numbers

Cloudian HyperStore achieves these breakthrough performance levels through a combination of architectural innovations:

Conclusion

As AI workloads continue to push the boundaries of infrastructure performance, storage systems must evolve to match the computational power of modern GPUs. Our testing demonstrates that Cloudian HyperStore is ready to meet this challenge, delivering exceptional object storage performance that scales linearly with infrastructure growth.

The full details of our testing methodology and results are available in the Cloudian-NVIDIA-Supermicro Full AI Stack Reference Architecture document, which provides a comprehensive blueprint for organizations looking to deploy high-performance AI infrastructure.

For more information on implementing Cloudian HyperStore in your AI infrastructure, visit www.cloudian.com.

Download the full stack reference architecture here.

Exit mobile version