Sunday, April 20, 2025
spot_img

ASUS Server Announces AI Performance Breakthrough With 26 Records in MLPerf Inference

spot_img
spot_img
spot_img
- Advertisement -
- Advertisement -

ASUS released its results for the first time since joining the MLCommons Association last December — instantly setting new performance records in dozens of benchmarked tasks.

Specifically, in the latest round of MLPerf Inference 2.0, ASUS servers set 26 records in the data center Closed division across six AI-benchmark tasks, outperforming all other servers with the same GPU configurations. The achievements consist of 12 records achieved with an ASUS ESC8000A-E11 server configured with eight 80 GB NVIDIA® A100 Tensor Core GPUs; and 14 records with an ASUS ESC4000A-E11 server with four 24 GB NVIDIA A30 Tensor Core GPUs.

These breakthrough results demonstrate clearly the performance dominance of ASUS servers in the AI arena — bringing significant value to organizations seeking to deploy AI and ensuring optimal performance in data centers.

ASUS set 26 records in AI inference, and dominates results tables across six tasks

The MLPerf Inference 2.0 benchmark covers six common AI-inferencing workloads, including image classification (ResNet50), object detection (SSD-ResNet34), medical image segmentation (3D-Unet), speech recognition (RNN-T), natural language processing (BERT) and recommendation (DLRM).

The dozen MLPerf Inference 2.0 12 records set by the NVIDIA-certified, 4U ESC8000A-E11 – configured with eight 80 GB NVIDIA A100 PCIe Tensor Core GPUs and two AMD EPYC 7763 CPUs – demonstrates its supreme scalability for AI and machine learning. Its streamlined thermal design, with independent CPU and GPU airflow tunnels, brings high-efficiency cooling solution to air-cooled data centers.

The NVIDIA-certified ESC4000A-E11, housed in the most compact 2U footprint on the market – and configured with four 24 GB NVIDIA A30 PCIe Tensor Core GPUs and two AMD EPYC 7763 CPUs – set a total of 14 MPLerf Inference 2.0 records. It offers a wide array of graphics accelerators, plus support for the NVIDIA NVLink high-speed GPU interconnect, to unleash maximum AI performance

If you have an interesting Article / Report/case study to share, please get in touch with us at editors@roymediative.com/ roy@roymediative.com, 9811346846/9625243429.

- Advertisement -
spot_img
spot_img
spot_img