Tag: bandwidth
-
Hacker News: AMD crafts custom EPYC CPU with HBM3 for Azure: 88 Zen 4 cores and 450GB of HBM3
Source URL: https://www.tomshardware.com/pc-components/cpus/amd-crafts-custom-epyc-cpu-for-microsoft-azure-with-hbm3-memory-cpu-with-88-zen-4-cores-and-450gb-of-hbm3-may-be-repurposed-mi300c-four-chips-hit-7-tb-s Source: Hacker News Title: AMD crafts custom EPYC CPU with HBM3 for Azure: 88 Zen 4 cores and 450GB of HBM3 Feedly Summary: Comments AI Summary and Description: Yes Summary: Microsoft has unveiled a new series of high-performance computing (HPC) Azure virtual machines, the HBv5 series, which utilize a custom AMD CPU…
-
The Register: Microsoft unveils beefy custom AMD chip to crunch HPC workloads on Azure
Source URL: https://www.theregister.com/2024/11/20/microsoft_azure_custom_amd/ Source: The Register Title: Microsoft unveils beefy custom AMD chip to crunch HPC workloads on Azure Feedly Summary: In-house DPU and HSM silicon also shown off Ignite One of the advantages of being a megacorp is that you can customize the silicon that underpins your infrastructure, as Microsoft is demonstrating at this…
-
The Cloudflare Blog: Bigger and badder: how DDoS attack sizes have evolved over the last decade
Source URL: https://blog.cloudflare.com/bigger-and-badder-how-ddos-attack-sizes-have-evolved-over-the-last-decade Source: The Cloudflare Blog Title: Bigger and badder: how DDoS attack sizes have evolved over the last decade Feedly Summary: If we plot the metrics associated with large DDoS attacks observed in the last 10 years, does it show a straight, steady increase in an exponential curve that keeps becoming steeper, or…
-
The Register: Nvidia’s latest Blackwell boards pack 4 GPUs, 2 Grace CPUs, and suck down 5.4 kW
Source URL: https://www.theregister.com/2024/11/18/nvidia_gb200_nvl4/ Source: The Register Title: Nvidia’s latest Blackwell boards pack 4 GPUs, 2 Grace CPUs, and suck down 5.4 kW Feedly Summary: You can now glue four H200 PCIe cards together too SC24 Nvidia’s latest HPC and AI chip is a massive single board computer packing four Blackwell GPUs, 144 Arm Neoverse cores,…
-
The Register: LLNL’s El Capitan surpasses Frontier with 1.74 exaFLOPS performance
Source URL: https://www.theregister.com/2024/11/18/top500_el_capitan/ Source: The Register Title: LLNL’s El Capitan surpasses Frontier with 1.74 exaFLOPS performance Feedly Summary: Uncle Sam tops supercomputer charts, while China recides from public view SC24 Lawrence Livermore National Lab’s (LLNL) El Capitan system has ended Frontier’s 2.5-year reign as the number one ranked supercomputer on the Top500, setting a new…
-
Cloud Blog: Data loading best practices for AI/ML inference on GKE
Source URL: https://cloud.google.com/blog/products/containers-kubernetes/improve-data-loading-times-for-ml-inference-apps-on-gke/ Source: Cloud Blog Title: Data loading best practices for AI/ML inference on GKE Feedly Summary: As AI models increase in sophistication, there’s increasingly large model data needed to serve them. Loading the models and weights along with necessary frameworks to serve them for inference can add seconds or even minutes of scaling…
-
The Register: Nvidia’s MLPerf submission shows B200 offers up to 2.2x training performance of H100
Source URL: https://www.theregister.com/2024/11/13/nvidia_b200_performance/ Source: The Register Title: Nvidia’s MLPerf submission shows B200 offers up to 2.2x training performance of H100 Feedly Summary: Is Huang leaving even more juice on the table by opting for mid-tier Blackwell part? Signs point to yes Analysis Nvidia offered the first look at how its upcoming Blackwell accelerators stack up…
-
Cloud Blog: Unlocking LLM training efficiency with Trillium — a performance analysis
Source URL: https://cloud.google.com/blog/products/compute/trillium-mlperf-41-training-benchmarks/ Source: Cloud Blog Title: Unlocking LLM training efficiency with Trillium — a performance analysis Feedly Summary: Rapidly evolving generative AI models place unprecedented demands on the performance and efficiency of hardware accelerators. Last month, we launched our sixth-generation Tensor Processing Unit (TPU), Trillium, to address the demands of next-generation models. Trillium is…