Supermicro last Thursday launched a full stack optimized storage solution for AI and ML data pipelines from data collection to high performance data delivery. This new solution maximizes AI time-to-value by keeping GPU data pipelines fully saturated. For AI training, massive amounts of raw data at petascale capacities can be collected, transformed, and loaded into an organization’s AI workflow pipeline. This multi-tiered Supermicro solution has been proven to deliver multi-petabyte data for AIOps and MLOps in production environments. The entire multi-rack scale solution from Supermicro is designed to reduce implementation risks, enable organizations to train models faster, and quickly use the resulting data for AI inference.

Petabytes of unstructured data used in large-scale AI training processing must be available to the GPU servers with low latencies and high bandwidth to keep the GPUs productive. Supermicro’s extensive portfolio of Intel and AMD based storage servers is a crucial element of the AI pipeline. These include the Supermicro Petascale All-Flash storage servers, which have a capacity of 983.04* TB per server of NVMe Gen 5 flash capacity and deliver up to 230 GB/s of read bandwidth and 30 million IOPS. This solution also includes the Supermicro SuperServer 90 drive bay storage servers for the capacity object tier. This complete and tested solution is available worldwide for customers in ML, GenAI, and other computationally complex workloads.The new storage solution consists of:
Switches: Supermicro InfiniBand and Ethernet Switches
All-Flash tier – Supermicro Petascale Storage Servers
Application tier – Supermicro 8U GPU Servers: AS -8125GS-TNHR and SYS-821GE-TNHR
Object tier – Supermicro 90 drive bay 4U SuperStorage Server running Quantum ActiveScale object storage
Software: WEKA Data Platform and Quantum ActiveScale object storage

Comments