Nvidia Announces "Super AI Chip" H200 Expected to Start Supply in the Second Quarter of Next Year
六月清晨搅
发表于 2023-11-14 18:36:59
305
0
0
Huang Renxun has upgraded his "equipment" again. On November 14th, a reporter from Daily Economic News learned from NVIDIA that on November 13th local time, NVIDIA announced the launch of NVIDIA HGX H200 (AI chip model, hereinafter referred to as "H200"). It is reported that H200 is the first to use HBM3e GPU (memory, faster and larger than before), further accelerating generative AI and large language models, while promoting scientific computing for HPC (high-performance computing) workloads. It can provide 141GB of display memory with a transmission speed of 4.8 TB/s, nearly doubling the capacity and bandwidth compared to the previous generation architecture of NVIDIA A100.
In the view of Ian Buck, Vice President of NVIDIA's Ultra Large Scale and High Performance Computing, in order to create intelligence through generative AI and HPC applications, it is necessary to use large, fast GPU graphics memory to process massive amounts of data quickly and efficiently. When H200 is used in conjunction with NVIDIA Grace CPUs using ultra fast NVLink C2C interconnect technology, it forms the GH200 Grace Hopper superchip with HBM3e - a computing module designed specifically for large-scale HPC and AI applications.
From the perspective of specifications, H200 will provide options for four and eight way H200 server motherboards, which are compatible with the hardware and software of the HGX H100 system; It can also be used for the NVIDIA GH200 Grace Hopper superchip using HBM3e released in August this year. These configurations enable H200 to be deployed in various data centers, including local, cloud, hybrid cloud, and edge; It can provide the highest performance for various application workloads, including LLM training and inference for super large models with parameters above 175 billion.
In terms of architecture, compared to the previous generation, the NVIDIA Hopper architecture has achieved a performance leap, such as nearly doubling the inference speed on a 70 billion parameter LLM - Llama 2 compared to the H100 (NVIDIA AI chip).
According to NVIDIA, H200 will be available through global system manufacturers and cloud service providers starting in the second quarter of 2024; Server manufacturers and cloud service providers are also expected to start offering systems equipped with H200 at the same time.
CandyLake.com 系信息发布平台,仅提供信息存储空间服务。
声明:该文观点仅代表作者本人,本文不代表CandyLake.com立场,且不构成建议,请谨慎对待。
声明:该文观点仅代表作者本人,本文不代表CandyLake.com立场,且不构成建议,请谨慎对待。
猜你喜欢
- Who are on the "Preferred Chip Stocks List" of Bank of America in 2025?
- Guo Mingchi: Apple M5 series chips will adopt TSMC N3P process, which entered the prototype stage several months ago
- Elon Musk's AI becomes Silicon Valley darling, $6 billion financing luxury lineup revealed, "old friends" such as Nvidia, AMD added
- Attraction crushing wide base index! Retail investors net purchase $29.8 billion worth of Nvidia stocks in 2024
- Nvidia New Product Countdown: New 'Nuclear Bomb' RTX 5090 Coming Soon, B300 Coming Soon
- Over 210 billion yuan in explosive purchases! Retail investors' fierce pursuit 'of Nvidia investment bank, optimistic about next year's performance
- NVIDIA's new 'nuclear bomb' leaked!
- NVIDIA's latest statement! Robot 'ChatGPT Moment' is Coming, Bet on the Next Growth Driver
- Nvidia may launch robot 'brain' in the first half of next year, with the company's stock price increasing by over 176% since the beginning of this year
- Nvidia plans to release a new generation of humanoid robot computing platform in the first half of next year, supporting multimodal AI models