DDN Storage Solutions Deliver 700% Gains in AI and Machine Learning for Image Segmentation and Natural Language Processing

DDN®, a leader in artificial intelligence (AI) and multi-cloud data management solutions, announced impressive performance results of its AI storage platform for the inaugural AI storage benchmarks released this week by MLCommons Association. The MLPerfTM Storage v0.5 benchmark results confirm DDN storage solutions as the gold standard for AI and machine learning applications.

DDN’s entries cover Image Segmentation and Natural Language Processing categories of the MLPerf Storage Benchmark. Using powerful single and multi-node GPU configurations, DDN’s A3I® AI400X2 storage appliance effortlessly scales to deliver faster and more reliable data access, while maximizing GPU utilization and delivering highest efficiency for demanding AI workloads.

In an individual compute node evaluation, a single DDN AI400X2 NVMe appliance equipped with DDN’s EXAScaler® 6.2 parallel filesystem fully served 40 AI accelerators, delivering a remarkable throughput of 16.2 GB/s.1 In a multi-node configuration, the same DDN AI400X2 NVMe appliance quadrupled its output, serving 160 accelerators over ten GPU compute nodes, achieving a throughput of 61.6 GB/s.2 These remarkable results demonstrate 700% better efficiency on a per storage node basis when compared to the competitive on-premises solution submissions.

“DDN’s cutting-edge data storage solutions fuel and accelerate GPUs in data centers and in the cloud, helping organizations develop better cancer detection methodologies, putting safe and reliable robotaxis on our roads and highways, and bringing to market more effective chatbots and virtual assistants to make our lives easier,” said Dr. James Coomer, SVP of Products at DDN. “We’re proud to lead the way in safe and power-efficient AI adoption, setting new standards for innovation and performance in the industry.”

The ability to power AI workloads, machine learning and Large Language Models at highest levels of efficiency and scale, while minimizing power usage and data center footprint is critical. With thousands of systems deployed on premise and in the cloud, DDN’s AI infrastructure storage systems are the solution of choice to power GPUs for the most demanding and innovative organizations in the world.

About MLCommons

MLCommons is an open engineering consortium with a mission to make machine learning better for everyone through benchmarks and data. The foundation for MLCommons began with the MLPerf benchmark in 2018, which rapidly scaled as a set of industry metrics to measure machine learning performance and promote transparency of machine learning techniques. In collaboration with its 50+ members – global technology providers, academics, and researchers, MLCommons is focused on collaborative engineering work that builds tools for the entire machine learning industry through benchmarks and metrics, public datasets, and best practices.

Sign up for the free insideAI News newsletter.

Join us on Twitter: https://twitter.com/InsideBigData1

Join us on LinkedIn: https://www.linkedin.com/company/insidebigdata/

Join us on Facebook: https://www.facebook.com/insideAI NewsNOW