It has taken untold thousands of people to make machine learning, and specifically the deep learning variety, the most viable form of artificial intelligence. …
The Buck Still Stops Here For GPU Compute was written by Timothy Prickett Morgan at The Next Platform.
There are no shortage of options for enterprises when it comes to Kubernetes platforms. …
Cisco Rolls Out New Systems As It Pushes Its Own Kubernetes Stack was written by Jeffrey Burt at The Next Platform.
Paid Feature Few organizations have the resources and talent on hand to skillfully navigate HPC infrastructure management and the emerging demands of AI/ML training and inference simultaneously. …
A Seamless Close To The HPC-AI Infrastructure Gap was written by David Gordon at The Next Platform.
Normally, when we look at a system, we think from the compute engines at a very fine detail and then work our way out across the intricacies of the nodes and then the interconnect and software stack that scales it across the nodes into a distributed computing platform. …
Nvidia Will Be A Prime Contractor For Big AI Supercomputers was written by Timothy Prickett Morgan at The Next Platform.
Nvidia and VMware have forged a tight partnership when it comes to bringing AI to the enterprise, which stands to reason given the prevalence of VMware’s ESXi hypervisor and vSphere management tools across more than 300,000 companies worldwide. …
Red Hat Stacks Up Software To Contain AI On Nvidia Platforms was written by Jeffrey Burt at The Next Platform.
When the “Aldebaran” datacenter GPUs were launched by AMD last November for the HPC and AI crowd pushing up into the exascale stratosphere, only the two top-end models of the Instinct GPU accelerators –ones that use the Open Accelerator Module (OAM) form factor put forth by Facebook and Microsoft under the Open Compute Project – were actually available. …
AMD Rounds Out “Aldebaran” GPU Lineup With Instinct MI210 was written by Timothy Prickett Morgan at The Next Platform.
There are a lot of things that compute engine makers have to do if they want to compete in the datacenter, but perhaps the most important thing is to be consistent. …
“Hopper” GH100 GPUs Are The Heart Of A More Expansive Nvidia System was written by Timothy Prickett Morgan at The Next Platform.
Hewlett Packard Enterprise was a bit ahead of the curve when it announced in 2019 that by this year it would be offering its entire product portfolio as a service and that its GreenLake hybrid cloud platform would be the foundation of the effort. …
HPE Focuses On Networking, HPC, Storage With GreenLake Expansion was written by Jeffrey Burt at The Next Platform.
Last fall ahead of the SC21 supercomputing conference, AMD said it was going to be the first of the major compute engine makers to add 3D vertical L3 cache to its chips, in this case to variants of the “Milan” Epyc 7003 series of processors that debuted in March 2021 called the “Milan-X” chips. …
“Milan-X” 3D Vertical Cache Yields Epyc HPC Bang For The Buck Boost was written by Timothy Prickett Morgan at The Next Platform.
Here we are, on the Friday before the flagship GPU Technology Conference hosted by Nvidia is set to kick off. …
Why Nvidia Should Acquire SUSE was written by Timothy Prickett Morgan at The Next Platform.
Google and VMware have announced a new element to their partnership that the two companies said will simplify cloud migrations, provide more flexibility, and help companies modernize their enterprise applications with a minimum amount of pain. …
VMware Widens Its Road To The Cloud With Google Partnership was written by Brandon Vigliarolo at The Next Platform.
Regular readers of The Next Platform know there are few more attention-grabbing tech events than Nvidia’s GPU Technology Conference (GTC). …
Edge. Smart Cities. Retail AI. Oh My! was written by David Gordon at The Next Platform.
To hardware or not to hardware, that is a real question for vendors peddling all kinds of software in the datacenter. …
Liqid Launches Its Own Systems, Chases AI And HPC was written by Timothy Prickett Morgan at The Next Platform.
It is not every day when we hear about a new supercomputer maker with a new architecture, but it is looking like Luminous Computing, a silicon photonics startup that has been pretty secretive about what it was up to, is going to be throwing its homegrown architecture into the ring. …
Luminous Shines A Light On Optical Architecture For Future AI Supercomputer was written by Jeffrey Burt at The Next Platform.
Paid Post There’s no doubt that the repurposing of GPU silicon has accelerated the development of artificial intelligence technology over the last decade. …
If You Want To Maximize Enterprise AI, Don’t Just Focus On GPUs was written by David Gordon at The Next Platform.
Creating a platform is a massive technical challenge. We have seen many technically elegant ones in recent years – Cloud Foundry, Engine Yard, the original OpenShift, Photon Platform, Mesos, OpenStack come immediately to mind – that didn’t quite make it, and importantly did not rise to the economic challenge of making enough money to sustain the continued development and support of that platform to have to reach tens of thousands, to hundreds of thousands, to millions of customers. …
Burning Cash Like Rocket Fuel To Get Hashi Stack To The Next Stage was written by Timothy Prickett Morgan at The Next Platform.
The Slurm Workload Manager that has its origins at Lawrence Livermore National Laboratory as the Simple Linux Utility for Resource Management – and which is used on about two-thirds of the most powerful HPC systems in the world – is looking for new jobs to take on across hybrid cloud infrastructure and machine learning systems running at scale. …
Slurm HPC Job Scheduler Applies For Work In AI And Hybrid Cloud was written by Daniel Robinson at The Next Platform.
When any new abstraction layer comes to compute, it can only think in integers at first, and then it learns to do fractions and finally, if we are lucky – and we are not always lucky – that abstraction layer learns to do multiplication and scale out across multiple nodes as well as scaling in – slicing itself into pieces – within a single node. …
Teaching Kubernetes To Do Fractions And Multiplication On GPUs was written by Timothy Prickett Morgan at The Next Platform.
Whenever demand exceeds supply, inflation is inevitable. And it is not at all surprising to find that in certain sectors of the networking space, the cost of bandwidth is flattening out instead of decreasing and in some cases is on the rise. …
Ethernet Switching Keeps Rising Despite Supply Chain Woes was written by Timothy Prickett Morgan at The Next Platform.
Over the past decade, much of the focus with machine learning has been on CPUs and accelerators, primarily GPUs but also custom ASICs, with advances in the chip architecture aimed at boosting parallel math performance. …
Architecting Memory Pools For HPC And AI Applications Using CXL was written by Jeffrey Burt at The Next Platform.