Tag: AI accelerators
-
Cloud Blog: Save on GPUs: Smarter autoscaling for your GKE inferencing workloads
Source URL: https://cloud.google.com/blog/products/containers-kubernetes/tuning-the-gke-hpa-to-run-inference-on-gpus/ Source: Cloud Blog Title: Save on GPUs: Smarter autoscaling for your GKE inferencing workloads Feedly Summary: While LLM models deliver immense value for an increasing number of use cases, running LLM inference workloads can be costly. If you’re taking advantage of the latest open models and infrastructure, autoscaling can help you optimize…
-
The Register: Fujitsu delivers GPU optimization tech it touts as a server-saver
Source URL: https://www.theregister.com/2024/10/23/fujitsu_gpu_middleware/ Source: The Register Title: Fujitsu delivers GPU optimization tech it touts as a server-saver Feedly Summary: Middleware aimed at softening the shortage of AI accelerators Fujitsu has started selling middleware that optimizes the use of GPUs, so that those lucky enough to own the scarce accelerators can be sure they’re always well-used.……
-
The Register: Uncle Sam reportedly considers capping AI chip shipments to Middle East
Source URL: https://www.theregister.com/2024/10/16/us_export_cap_ai_chip_middle_east/ Source: The Register Title: Uncle Sam reportedly considers capping AI chip shipments to Middle East Feedly Summary: Meanwhile, ASML shares plunge as China export curbs hit forecast Amid fears that Middle Eastern nations are being used to funnel AI accelerators and other sensitive technologies to China, the Biden administration is reportedly considering…
-
Cloud Blog: Sustainable silicon to intelligent clouds: collaborating for the future of computing
Source URL: https://cloud.google.com/blog/topics/systems/2024-ocp-global-summit-keynote/ Source: Cloud Blog Title: Sustainable silicon to intelligent clouds: collaborating for the future of computing Feedly Summary: Editor’s note: Today, we hear from Parthasarathy Ranganathan, Google VP and Technical Fellow and Amber Huffman, Principal Engineer. Partha delivered a keynote address today at the 2024 OCP Global Summit, an annual conference for leaders,…
-
The Register: AMD targets Nvidia H200 with 256GB MI325X AI chips, zippier MI355X due in H2 2025
Source URL: https://www.theregister.com/2024/10/10/amd_mi325x_ai_gpu/ Source: The Register Title: AMD targets Nvidia H200 with 256GB MI325X AI chips, zippier MI355X due in H2 2025 Feedly Summary: Less VRAM than promised, but still gobs more than Hopper AMD boosted the VRAM on its Instinct accelerators to 256 GB of HBM3e with the launch of its next-gen MI325X AI…
-
The Register: LiquidStack says its new CDU can chill more than 1MW of AI compute
Source URL: https://www.theregister.com/2024/08/22/liquidstack_cdu_ai/ Source: The Register Title: LiquidStack says its new CDU can chill more than 1MW of AI compute Feedly Summary: So what’s that good for? Like eight of Nvidia’s NVL-72s? As GPUs and AI accelerators push beyond one kilowatt of power consumption, many systems builders are turning to liquid cooling to manage the…