Xinnor Blog
Author Image
WRITTEN BY Celestica

2026-03-10

In the high-stakes world of Artificial Intelligence and High-Performance Computing (HPC), storage is often the "silent bottleneck." While GPUs get the attention for processing massive datasets, they are only as effective as the storage infrastructure feeding them. At the Friedrich-Alexander-Universität Erlangen-Nürnberg (FAU), researchers recently faced this challenge head-on when deploying Helma, the most powerful university-owned AI supercomputer in Germany.

To unlock the full potential of Helma’s 768 NVIDIA H100/H200 GPUs, NHR@FAU required a storage solution that could handle terabyte-scale training sets and frequent model checkpoints with ease. The answer lay in a collaborative design featuring Celestica storage controllers.

The Foundation of Performance: Celestica SC6100

SC6100 Branded
Celestica's SC6100


To meet NHR@FAU’s non-negotiable requirements for high availability and top-tier  performance, MEGWARE (a system integrator specializing in HPC and AI infrastructure) and Xinnor (a software company that specializes in patented software RAID technology) selected the Celestica SC6100. The SC6100 is a high-availability, all-flash PCIe Gen5 storage controller designed specifically for demanding workloads like AI/ML and atomistic simulation.

In the Helma configuration, ten (10) SC6100 controllers form the backbone of a half-rack, all-NVMe Lustre file system. Each controller is a powerhouse, featuring:

  • Dual-Node Redundancy: Two independent server nodes sharing 24 PCIe Gen5 NVMe SSDs and three PCIe Gen5 x16 slots for high speed network IO, ensuring the system continues to operate even if a node fails.

  • High-Speed Processing: Powered by AMD EPYC 9454P CPUs (48 cores) and 384 GB of DDR5 RAM per server node to handle massive metadata and throughput demands. (SC6100 can be configured with a range of AMD EPYC CPUs up to 64 cores and memory up to 1.5TB RAM - per server node

Achieving "Podium" Performance

The results of this architecture speak for themselves. After deployment, the Helma cluster was  submitted to IO500, the premier global ranking for high-performance storage. The Celestica-powered solution achieved:

  • #1 position among all Lustre-based solutions.

  • #3 ranking in the global IO500 benchmark.

  • 1,798.77 GiB/s in IOR Easy Read performance.

  • Over 8.2 Million IOPS in metadata testing (MDtest Easy Stat).

Perhaps most impressively, Celestica helped deliver this world-class performance within a half-rack footprint, roughly one-tenth the size of many competing multi-rack appliances.  This allows not just for less physical space, but less associated power and cooling along with lower hardware and software costs.

“For a national supercomputing center like NHR@FAU, performance is vital, but reliability is non-negotiable. Our collaboration with Celestica and MEGWARE allowed us to build a seamless, high-availability architecture where xiRAID and Lustre work in lockstep with the SC6100’s dual-node redundancy. This setup ensures that whether it’s a drive failure or a complete node failure, the system remains online, protecting the massive datasets and long-running GPU training sessions that are critical to Germany’s AI research.” -- Dmitry Livshits, CEO, Xinnor

Why Celestica Matters for AI Infrastructure

Celestica’s contribution to the Helma supercomputer extends beyond raw speed. By utilizing off-the-shelf commodity hardware, Celestica enables NHR@FAU to avoid vendor lock-in and benefit from "commodity economics".

The SC6100’s integration with Xinnor’s xiRAID and Lustre creates a "Green Efficiency" effect. By maximizing GPU utilization with fewer physical servers, the university reduces power and cooling OPEX expenditures while accelerating innovation in fields like quantum chemistry and biomedical image analysis.

Looking Ahead

With the Helma supercomputer now live, NHR@FAU is equipped to support the next generation of AI research. The inclusion of PCIe Gen 5 NVMe within the Celestica controllers ensures the system is future-proof, leaving significant headroom for future expansions as AI models continue to grow in complexity.  Moreover, MEGWARE and Xinnor are moving forward with similar high-performance storage solutions using the SC6100 for other customers with demanding HPC and AI workloads.

The Helma project proves that with the right hardware partner, academic institutions can achieve "exa-scale" performance without the "exa-scale" footprint or budget.

To learn more, and dive into the details of this solution, read the full case-study