CoreWeave, a leading AI Hyperscaler, has become the first cloud provider to bring NVIDIA’s H200 Tensor Core GPUs to market, marking a significant milestone in the development of artificial intelligence infrastructure. This innovative platform combines NVIDIA’s cutting-edge GPUs with Intel’s fifth-generation Xeon CPUs and 3200Gbps of NVIDIA Quantum-2 InfiniBand networking, enabling customers to train large-scale generative AI models with unprecedented efficiency.
The H200 GPU boasts impressive specs, including 4.8 TB/s memory bandwidth and 141 GB GPU memory capacity, delivering up to 1.9X higher inference performance than its predecessor, the H100 GPU. Michael Intrator, CEO and co-founder of CoreWeave, praised the collaboration with NVIDIA, stating that this technology empowers customers to tackle complex AI models with unprecedented efficiency. Ian Buck, vice president of Hyperscale and HPC at NVIDIA, also commended CoreWeave’s ability to rapidly deploy NVIDIA technology, ensuring customers have access to cutting-edge tools for training large language models.
Accelerating AI Infrastructure Performance with NVIDIA H200 Tensor Core GPUs
CoreWeave, a leading cloud provider, has announced that it is the first to market with NVIDIA H200 Tensor Core GPUs, ushering in a new era of AI infrastructure performance. This innovative platform delivers performant AI infrastructure with high system reliability and resilience, enabling customers to use NVIDIA H200 GPUs at scale to accelerate the development of their generative AI applications.
The NVIDIA H200 Tensor Core GPU is designed to push the boundaries of generative AI by providing 4.8 TB/s memory bandwidth and 141 GB GPU memory capacity, which helps deliver up to 1.9X higher inference performance than H100 GPUs. CoreWeave’s H200 instances combine NVIDIA H200 GPUs with Intel’s fifth-generation Xeon CPUs (Emerald Rapids) and 3200Gbps of NVIDIA Quantum-2 InfiniBand networking, and are deployed in clusters with up to 42,000 GPUs and accelerated storage solutions to deliver powerful performance.
CoreWeave’s Mission Control platform offers customers unmatched reliability and resiliency by managing the complexities of AI infrastructure deployment and uptime with software automation. The platform helps customers train models faster and more efficiently by using advanced system validation processes, proactive fleet health-checking, and extensive monitoring capabilities. This translates to reduced system downtime, faster time to solution, and lower total cost of ownership.
Unparalleled Performance for Generative AI Applications
The combination of H200 GPUs with CoreWeave’s technology empowers customers to tackle the most complex AI models with unprecedented efficiency and achieve new levels of performance. With NVIDIA H200 GPUs, customers can accelerate the development of their generative AI applications, enabling them to train larger and more complex models faster and more efficiently.
CoreWeave’s infrastructure services are used to train some of the largest and most ambitious models from customers including Cohere, Mistral, and NovelAI. The company’s proven track record of being first to market with large-scale AI infrastructure has enabled it to deliver a large-scale NVIDIA H100 Tensor Core GPU cluster interconnected with NVIDIA Quantum-2 InfiniBand networking, which broke MLPerf training records in June 2023.
Advanced System Validation and Monitoring Capabilities
CoreWeave’s Mission Control platform provides customers with advanced system validation processes, proactive fleet health-checking, and extensive monitoring capabilities. This enables customers to maintain uninterrupted AI development pipelines, reducing system downtime and lowering total cost of ownership. The platform’s rich suite of observability tools and services provides transparency across all critical components of the system, empowering teams to optimize their AI infrastructure for maximum performance.
Rapid Scaling of Data Center Operations
In addition to bringing the latest NVIDIA GPUs to market and advancing its portfolio of cloud services, CoreWeave is rapidly scaling its data center operations to keep up with demand for its industry-leading infrastructure services. The company has completed nine new data center builds since the beginning of 2024, with 11 more in progress. CoreWeave expects to end the year with 28 data centers globally, with an additional 10 new data centers planned in 2025.
Industry-Leading Infrastructure Services
CoreWeave’s technology provides enterprises and leading AI labs with the most performant and efficient cloud solutions for accelerated computing. Since 2017, CoreWeave has operated a growing footprint of data centers covering every region of the US and across Europe. The company was ranked as one of the TIME100 most influential companies and featured on Forbes Cloud 100 ranking in 2024.
External Link: Click Here For More
