News from the AI & ML world
Andrew Jolly@AIwire
//
NVIDIA is making significant strides in AI computing with the release of RTX AI PCs and Workstations, designed to accelerate coding assistant performance. These AI-powered copilots are fundamentally changing software development, providing real-time assistance to both experienced and novice developers. Coding assistants, optimized for RTX AI PCs, offer suggestions, explanations, and debugging capabilities, streamlining tasks and enhancing productivity across various projects, from academic endeavors to production code. These assistants can run locally, eliminating the latency and subscription costs associated with cloud-based alternatives.
CoreWeave has emerged as the first cloud platform to offer NVIDIA RTX PRO 6000 Blackwell Server Edition instances at scale. This advancement provides users with significantly improved performance, achieving up to 5.6x faster LLM inference and 3.5x faster text-to-video generation compared to previous generations. The RTX PRO 6000 is tailored for inference of models up to 70B parameters, providing a cost-efficient alternative to larger GPU clusters while maintaining strong performance for teams developing and scaling AI applications. CoreWeave now boasts one of the widest ranges of NVIDIA Blackwell infrastructure on the market, which also includes the NVIDIA GB200 NVL72 system and NVIDIA HGX B200 platform.
Amazon Web Services (AWS) has also announced the general availability of EC2 P6e-GB200 UltraServers, powered by NVIDIA Grace Blackwell GB200 superchips. These UltraServers deliver up to 72 GPUs with 360 petaflops of computing power, catering to AI training and inference at the trillion-parameter scale. The NVIDIA Grace Blackwell Superchips integrate two high-performance NVIDIA Blackwell tensor core GPUs and an NVIDIA Grace CPU, connected by the NVIDIA NVLink-C2C interconnect, boosting bandwidth between the GPU and CPU significantly. These UltraServers are deployed in EC2 UltraClusters, providing secure and reliable scalability to tens of thousands of GPUs, and are ideal for compute-intensive AI workloads such as training frontier models and building generative AI applications.
ImgSrc: www.aiwire.net
References :
- AWS News Blog: Amazon announces the general availability of EC2 P6e-GB200 UltraServers, powered by NVIDIA Grace Blackwell GB200 superchips that enable up to 72 GPUs with 360 petaflops of computing power for AI training and inference at the trillion-parameter scale.
- AIwire: CoreWeave, Inc. today announced it is the first cloud platform to make NVIDIA RTX PRO 6000 Blackwell Server Edition instances generally available.
Classification: