AMD launches new AI high-performance computing solution
因醉鞭名马幌
发表于 2024-10-12 10:17:11
159
0
0
On October 11th, at Advancing AI 2024, AMD launched a new AI high-performance computing solution, including the fifth generation AMD EPYC server CPU, AMD Instinct MI325X accelerator, AMD Pensando Salina DPU, AMD Pensando Pollara 400 NIC, and AMD Ryzen AI PRO 300 series processor for enterprise AI PCs.
Dr. Su Zifeng, Chairman and CEO of AMD, stated, "With our new EPYC CPU, AMD Instinct GPU, and Pensando DPU, we will provide leading computing power to support our customers' most important and demanding workloads. Looking ahead, we expect the market size of data center AI accelerators to grow to $500 billion by 2028
In 2018, AMD EPYC server CPUs only had a market share of 2%. In less than 7 years, it has now reached 34%. Data centers and AI have brought huge growth opportunities for AMD.
EPYC CPU fully upgraded
As one of AMD's core products, AMD EPYC server CPUs have undergone a comprehensive upgrade.
The fifth generation AMD EPYC server CPU (AMD EPYC 9005 series CPU), codenamed "Turin", adopts the "Zen5" core architecture, is compatible with the SP5 platform, provides up to 192 cores, and has a maximum frequency of 5GHz. The AVX512 instruction set supports a complete 512 bit wide data path.
AMD states that the AMD EPYC 9005 is an advanced CPU designed for AI. Compared with traditional hardware, AMD EPYC 9005 can achieve equivalent integer computing performance while significantly reducing the number of racks, greatly reducing the physical space, power consumption, and required software license quantity, thus freeing up space for new or growing AI workloads.
AMD also stated that the AMD EPYC 9005 has outstanding AI inference performance. Compared to the previous generation product, servers running two 5th generation AMD EPYC 9965 CPUs can provide up to twice the inference throughput capability.
AMD has proven that it can meet the needs of the data center market and set a benchmark for data center performance, efficiency, solutions, and functionality that can meet the demands of cloud, enterprise, and AI workloads for customers, "said Dan McNamara, Senior Vice President and General Manager of AMD's Server Division
Instinct GPU steadily advances
As an important carrier of AI computing power, AMD Instinct GPU has also undergone updates and iterations. In addition, AMD has also announced its GPU product roadmap for 2025 and 2026.
The AMD Instinct MI325X is built on the third-generation AMD CDNA architecture, featuring 256GB of HBM3E memory and 6TB/s of memory bandwidth, delivering impressive training and inference performance and efficiency, setting a new standard for AI performance. According to data released by AMD, the AMD Instinct MI325X outperforms the Nvidia H200 in inference across multiple models.
AMD stated that the AMD Instinct MI325X is expected to be put into production and shipped in the fourth quarter of 2024, while the complete system and infrastructure solutions of partners such as Dell, Gigabyte, HP, and Lenovo will be launched from the first quarter of 2025.
In terms of future product layout, compared to accelerators based on AMD CDNA 3 architecture, the inference performance of AMD Instinct MI350 based on AMD CDNA 4 architecture will be improved by 35 times. Meanwhile, the AMD Instinct MI350 can be equipped with up to 288GB of HBM3E memory and is expected to be launched in the second half of 2025.
AMD also announced significant progress in the development of the AMD Instinct MI400 based on the AMD CDNA Next architecture, with plans to launch it in 2026.
Improve the performance of AI networks
Currently, AI networks are crucial for ensuring effective utilization of CPUs and accelerators in AI infrastructure.
To support the next generation of AI networks, AMD is utilizing widely deployed programmable DPUs to provide support for ultra large scale computing. The AI network can be divided into two parts: the front-end that transmits data and information to the AI cluster, and the back-end that manages data transmission between the accelerator and the cluster. To this end, AMD has launched the AMD Pensando Salina DPU for front-end and the AMD Pensando Pollara 400 for back-end.
The AMD Pensando Salina DPU is one of the world's highest performing and most programmable third-generation DPUs, with twice the performance, bandwidth, and scale compared to its predecessor. The AMD Pensando Salina DPU supports a throughput of 400G and enables high-speed data transmission. It is a key component in AI front-end networks, optimizing performance, efficiency, security, and scalability for data-driven AI applications.
The Pensando Pollara 400 is the industry's first UEC ready AI NIC (an AI network card that complies with the Super Ethernet Alliance specifications), which supports next-generation RDMA software and an open network ecosystem, ensuring performance leading, scalable, and efficient communication between accelerators in the backend network.
In terms of launch time, AMD Pensando Salina DPU and AMD Pensando Pollara 400 will both provide samples to customers in the fourth quarter of 2024 and are expected to be launched in the first half of 2025.
Forrest Norrod, Executive Vice President and General Manager of AMD's Data Center Solutions Division, said, "With the new AMD Instinct Accelerator, EPYC Processor, AMD Pensando Network Engine, open software ecosystem, and the ability to integrate these things into AI infrastructure, AMD fully possesses the key expertise to build and deploy world-class AI solutions
CandyLake.com is an information publishing platform and only provides information storage space services.
Disclaimer: The views expressed in this article are those of the author only, this article does not represent the position of CandyLake.com, and does not constitute advice, please treat with caution.
Disclaimer: The views expressed in this article are those of the author only, this article does not represent the position of CandyLake.com, and does not constitute advice, please treat with caution.
You may like
- Bit Digital completes acquisition of Enovum data center, expanding high-performance computing business
- Intel、AMD联合捍卫x86!黄仁勋:他们很了不起
- Intel and AMD join forces to defend x86! Huang Renxun: They are amazing
- Intel, AMD 통합 x86!황인훈: 걔네들 못 살아.
- 以色列财政部下调该国今明两年经济增长预期
- AMD携端到端AI解决方案“出场”
- AMD brings end-to-end AI solutions to the market
- AMD、エンドツーエンドのAIソリューションを「登場」
- AMD 휴대용 종단간 AI 솔루션'출전'
-
세계 최대 호텔 그룹인 메리어트 인터내셔널 그룹이 기업 본사가 있는 메릴랜드주 정부 홈페이지에 제출한 정보에 따르면 2025년 초 833명을 감원할 계획이라는 보도가 나왔다.메리어트가 지난해 대외적으로 언급한 ...
- snail_007
- 어제 22:38
- Up
- Down
- Reply
- Favorite
-
"장충모는 일찍 황인훈을 대적전 CEO로 초청한지 10분만에 거절당했다."11월 29일, 올해 93세의 대적전 창시자 장충모의 자서전 신권이 정식으로 출판되였다. 자서전에서 장충모는 엔비디아 창시자 CEO 황인훈과의 ...
- 茉莉707
- 어제 20:49
- Up
- Down
- Reply
- Favorite
-
"금리 인상이 다가오고 있다!일본 중앙은행은 방금 메가톤급 신호를 보냈다!"닛케이신문에 따르면 일본 중앙은행 총재 시다와 남자는 인플레이션과 경제 추세의 발전이 중앙은행의 예측에 부합함에 따라 금리 인상 ...
- 奶王钟浩斌
- 2 시간전
- Up
- Down
- Reply
- Favorite
-
베이징상보소식 (기자 류효몽): 현지시간으로 12월 1일, Stellantis그룹은 당유실 (Carlos Tavares) 이 Stellantis그룹 최고경영자직무를 사직할데 대한 청구를 접수하고 이 사표는 즉시 효력을 발생한다고 선포했 ...
- 遥远的梦
- 5 시간전
- Up
- Down
- Reply
- Favorite