Intel Launches Habana Gaudi4 Custom AI Accelerators for Large-Scale Model Training
Intel has unveiled the 'Habana Gaudi4,' a new custom AI accelerator designed to enhance large-scale model training and inference. The Gaudi4 features optimized memory paths and scalable multi-chip interconnects, targeting hyperscalers and enterprise AI teams. It offers higher throughput, improved tensor cores, and native support for major ML frameworks, integrating with Intel's data center stack. This launch aims to reduce infrastructure costs for organizations using large language models and reflects a trend towards diversified AI hardware.

Intel has launched the 'Habana Gaudi4,' a custom AI accelerator aimed at speeding up large-scale model training and inference. The Gaudi4 includes optimized memory paths and scalable multi-chip interconnects, catering to hyperscalers and enterprise AI teams.
It enhances previous Habana designs with increased throughput, improved tensor cores, and support for major ML frameworks. The chips integrate with Intel’s data center stack, and partnerships with cloud providers are established for early access and benchmarking of transformer workloads. Gaudi4 aims to lower infrastructure costs for organizations operating large language models and highlights a shift towards diverse AI hardware solutions.




Comments