When the “Aldebaran” datacenter GPUs were launched by AMD last November for the HPC and AI crowd pushing up into the exascale stratosphere, only the two top-end models of the Instinct GPU accelerators –ones that use the Open Accelerator Module (OAM) form factor put forth by Facebook and Microsoft under the Open Compute Project – were actually available. …
AMD Rounds Out “Aldebaran” GPU Lineup With Instinct MI210 was written by Timothy Prickett Morgan at The Next Platform.
There are a lot of things that compute engine makers have to do if they want to compete in the datacenter, but perhaps the most important thing is to be consistent. …
“Hopper” GH100 GPUs Are The Heart Of A More Expansive Nvidia System was written by Timothy Prickett Morgan at The Next Platform.
Hewlett Packard Enterprise was a bit ahead of the curve when it announced in 2019 that by this year it would be offering its entire product portfolio as a service and that its GreenLake hybrid cloud platform would be the foundation of the effort. …
HPE Focuses On Networking, HPC, Storage With GreenLake Expansion was written by Jeffrey Burt at The Next Platform.
Last fall ahead of the SC21 supercomputing conference, AMD said it was going to be the first of the major compute engine makers to add 3D vertical L3 cache to its chips, in this case to variants of the “Milan” Epyc 7003 series of processors that debuted in March 2021 called the “Milan-X” chips. …
“Milan-X” 3D Vertical Cache Yields Epyc HPC Bang For The Buck Boost was written by Timothy Prickett Morgan at The Next Platform.
Here we are, on the Friday before the flagship GPU Technology Conference hosted by Nvidia is set to kick off. …
Why Nvidia Should Acquire SUSE was written by Timothy Prickett Morgan at The Next Platform.
Google and VMware have announced a new element to their partnership that the two companies said will simplify cloud migrations, provide more flexibility, and help companies modernize their enterprise applications with a minimum amount of pain. …
VMware Widens Its Road To The Cloud With Google Partnership was written by Brandon Vigliarolo at The Next Platform.
Regular readers of The Next Platform know there are few more attention-grabbing tech events than Nvidia’s GPU Technology Conference (GTC). …
Edge. Smart Cities. Retail AI. Oh My! was written by David Gordon at The Next Platform.
To hardware or not to hardware, that is a real question for vendors peddling all kinds of software in the datacenter. …
Liqid Launches Its Own Systems, Chases AI And HPC was written by Timothy Prickett Morgan at The Next Platform.
It is not every day when we hear about a new supercomputer maker with a new architecture, but it is looking like Luminous Computing, a silicon photonics startup that has been pretty secretive about what it was up to, is going to be throwing its homegrown architecture into the ring. …
Luminous Shines A Light On Optical Architecture For Future AI Supercomputer was written by Jeffrey Burt at The Next Platform.
Paid Post There’s no doubt that the repurposing of GPU silicon has accelerated the development of artificial intelligence technology over the last decade. …
If You Want To Maximize Enterprise AI, Don’t Just Focus On GPUs was written by David Gordon at The Next Platform.
Creating a platform is a massive technical challenge. We have seen many technically elegant ones in recent years – Cloud Foundry, Engine Yard, the original OpenShift, Photon Platform, Mesos, OpenStack come immediately to mind – that didn’t quite make it, and importantly did not rise to the economic challenge of making enough money to sustain the continued development and support of that platform to have to reach tens of thousands, to hundreds of thousands, to millions of customers. …
Burning Cash Like Rocket Fuel To Get Hashi Stack To The Next Stage was written by Timothy Prickett Morgan at The Next Platform.
The Slurm Workload Manager that has its origins at Lawrence Livermore National Laboratory as the Simple Linux Utility for Resource Management – and which is used on about two-thirds of the most powerful HPC systems in the world – is looking for new jobs to take on across hybrid cloud infrastructure and machine learning systems running at scale. …
Slurm HPC Job Scheduler Applies For Work In AI And Hybrid Cloud was written by Daniel Robinson at The Next Platform.
When any new abstraction layer comes to compute, it can only think in integers at first, and then it learns to do fractions and finally, if we are lucky – and we are not always lucky – that abstraction layer learns to do multiplication and scale out across multiple nodes as well as scaling in – slicing itself into pieces – within a single node. …
Teaching Kubernetes To Do Fractions And Multiplication On GPUs was written by Timothy Prickett Morgan at The Next Platform.
Whenever demand exceeds supply, inflation is inevitable. And it is not at all surprising to find that in certain sectors of the networking space, the cost of bandwidth is flattening out instead of decreasing and in some cases is on the rise. …
Ethernet Switching Keeps Rising Despite Supply Chain Woes was written by Timothy Prickett Morgan at The Next Platform.
Over the past decade, much of the focus with machine learning has been on CPUs and accelerators, primarily GPUs but also custom ASICs, with advances in the chip architecture aimed at boosting parallel math performance. …
Architecting Memory Pools For HPC And AI Applications Using CXL was written by Jeffrey Burt at The Next Platform.
If you need any proof that it doesn’t take the most advanced chip manufacturing processes to create an exascale-class supercomputer, you need look no further than the Sunway “OceanLight” system housed at the National Supercomputing Center in Wuxi, China. …
How China Made An Exascale Supercomputer Out Of Old 14 Nanometer Tech was written by Timothy Prickett Morgan at The Next Platform.
There are people who build big machines and then there are people who create the algorithms, libraries, and applications that harness them. …
Looking For A Singularity Event For Scientific Computing was written by Jeffrey Burt at The Next Platform.
There is a lot of chatter this week about optical communications with the OFC2022 conference being held in San Diego. …
Talking Silicon Photonics Signal And Noise With Andy Bechtolsheim was written by Timothy Prickett Morgan at The Next Platform.
Some of the most important luminaries in the HPC sector have spoken from on high, and their conclusions about the future of the HPC market are probably going to shock a lot of people. …
Will HPC Be Eaten By Hyperscalers And Clouds? was written by Timothy Prickett Morgan at The Next Platform.
When it comes to memory for compute engines, FPGAs – or rather what we have started calling hybrid FPGAs because they have all kinds of hard coded logic as well as the FPGA programmable logic on a single package – have the broadest selection of memory types of any kind of device out there. …
A Cornucopia Of Memory And Bandwidth In The Agilex-M FPGA was written by Timothy Prickett Morgan at The Next Platform.