Not all of the new and interesting high performance computing systems are always in the upper echelons of the Top 500 supercomputing list, which was announced at the opening of the SC16 supercomputing conference in Salt Lake City this week. Sometimes, an intriguing system breaks into the list outside of the top ten or twenty most powerful machines in the bi-annual rankings of number-crunching performance, and such is the case with the new “Saturn V” supercomputer built by Nvidia using its latest GPUs and interconnects.
The Saturn V system, nick-named of course for the NASA launch vehicle that eventually …
How Nvidia’s Own Saturn V DGX-1 Cluster Stacks Up was written by Timothy Prickett Morgan at The Next Platform.
The bi-annual rankings of the Top 500 supercomputers for the November 2016 systems are now live. While the top of the list is static with the same two Chinese supercomputers dominating, there are several new machines that have cropped up to replace decommissioned systems throughout, the momentum at the very top shows some telling architectural trends, particularly among the newcomers in the top 20.
We already described the status of the major Chinese and Japanese systems in our analysis of the June 2016 list and thought it might be more useful to look at some of the broader …
A Closer Look at 2016 Top 500 Supercomputer Rankings was written by Nicole Hemsoth at The Next Platform.
The latest listing of the Top 500 rankings of the world’s most powerful supercomputers has just been released. While there were no big surprises at the top of the list, there have been some notable additions to the top tier, all of which feature various elements of supercomputers yet to come as national labs and research centers prepare for their pre-exascale and eventual exascale systems.
We will be providing a deep dive on the list results this morning, but for now, what is most interesting about the list is what it is just beginning to contain at the top–and what …
Inside Six of the Newest Top 20 Supercomputers was written by Nicole Hemsoth at The Next Platform.
While the machine learning applications created by hyperscalers and the simulations and models run by HPC centers are very different animals, the kinds of hardware that help accelerate the performance for one is also helping to boost the other in many cases. And that means that the total addressable market for systems like the latest GPU-accelerated Power Systems machines or the alternatives from Nvidia and others has rapidly expanded as enterprises try to deploy both HPC and AI to better run their businesses.
HPC as we know it has obviously been around for a long time, and is in a …
IBM Shows Off AI And HPC Oomph On Power8 Tesla Hybrids was written by Timothy Prickett Morgan at The Next Platform.
Over the course of the last five years, GPU computing has featured prominently in supercomputing as an accelerator on some of the world’s fastest machines. If some supercomputer makers are correct, GPUs will continue to play a major role in high performance computing, but the acceleration they provide will go beyond boosts to numerical simulations. This has been great news for Nvidia’s bottom line since the market for GPU computing is swelling, and for HPC vendors that can integrate those and wrap the proper software stacks around both HPC and machine learning, it could be an equal boon. …
Cray’s New Pascal XC50 Supercomputer Points to Richer HPC Future was written by Nicole Hemsoth at The Next Platform.
An event as large and diverse as the annual Supercomputing Conference (SC16) presents a daunting array of content, even for those who specialize in a particular area inside the wider HPC spectrum. For HPC programmers, there are many sub-tracks to follow depending where on the stack on sits.
The conference program includes a “Programming Systems” label for easily finding additional relevant sessions, but we wanted to highlight a few of these here based on larger significance to the overall HPC programming ecosystem.
HPC programmers often have special considerations in how they program that other fields do not. For example, nothing …
SC16 for HPC Programmers: What to Watch was written by Nicole Hemsoth at The Next Platform.
Chip maker Nvidia was founded by people who loved gaming and who wanted to make better 3D graphics cards, and decades later, the company has become a force in computing, first in HPC and then in machine learning and now database acceleration. And it all works together, with gaming graphics providing the foundation on which Nvidia can build a considerable compute business, much as Intel’s PC business provided the foundation for its Xeon assault on the datacenter over the past two and a half decades.
At some point, Nvidia may not need an explicit link to PC graphics and gaming …
Pascal GPUs On All Fronts Push Nvidia To New Highs was written by Timothy Prickett Morgan at The Next Platform.
Burst buffer technology is closely associated with HPC applications and supercomputer sites as a means of ensuring that persistent storage, typically a parallel file system, does not become a bottleneck to overall performance, specifically where checkpoints and restarts are concerned. But attention is now turning to how burst buffers might find broader use cases beyond this niche, and how they could be used for accelerating performance in other areas where the ability to handle a substantial volume of data with high speed and low latency is key.
The term burst buffer is applied to this storage technology simply because this …
What Sort of Burst Buffer Are You? was written by Nicole Hemsoth at The Next Platform.
Moore’s Law may be slowing down performance increases in compute capacity, but InfiniBand networking did not get the memo. Mellanox Technologies has actually picked up the pace, in fact, and is previewing 200 Gb/sec InfiniBand switches and server adapters that are timed to come to market with a slew of Xeon, Opteron, ARM, and Power processors due around the middle of next year.
The new Quantum InfiniBand switch ASIC and its companion ConnextX-6 adapter ASICs come relatively hot on the heels of the 100 Gb/sec Enhanced Data Rate, or EDR, products that were announced in the fall of 2014 and …
InfiniBand Breaks Through The 200G Barrier was written by Timothy Prickett Morgan at The Next Platform.
The Java Virtual Machine (JVM) is a vital part of modern distributed computing. It is the platform of big data applications like Spark, HDFS, Cassandra,and Hive. While the JVM provides “write once, run anywhere” platform independence, this comes at a cost. The JVM takes time to “warm up”, that is to load the classes, interpret the bytecode, and so on. This time may not matter much for a long-running Tomcat server, but big data jobs are typically short-lived. Thus the parallelization often used to speed up the time-to-results compounds the JVM warmup time problem.
David Lion and his colleagues examined …
JVM Boost Shows Warm Java is Better than Cold was written by Nicole Hemsoth at The Next Platform.
There is no question that information technology is always too complex, and that people have been complaining about this for over five decades now. It keeps us employed, so perhaps we should not point this out, and moreover, perhaps we should not be so eager to automate ourselves out of jobs. But if the advance of computing from mainframes to artificial intelligence teach us anything, it is that we always want to make IT simpler to get people out of the way of doing business or research.
The founders of hyperconverged systems maker Nutanix learned its lessons from hyperscalers like …
Getting Hyper About Converged Storage, And Then Some was written by Timothy Prickett Morgan at The Next Platform.
Collecting data is only useful to the extent that the data is analyzed. These days, human Internet usage is generating more data (particularly for advertising purposes) and Internet of Things devices are providing data about our homes, our cars, and our bodies.
Analyzing that data can become a challenge at scale. Streaming platforms work well with incoming data but aren’t designed for post hoc analysis. Traditional database management systems can perform complex queries against stored data, but cannot be put to real-time usage.
One proposal to address these challenges, called Quill, was developed by Badrish Chandramouli and colleagues at Microsoft …
Microsoft Research Pens Quill for Data Intensive Analysis was written by Nicole Hemsoth at The Next Platform.
In the high performance computing arena, the stress is always on performance. Anything and everything that can be done to try to make data retrieval and processing faster ultimately adds up to better simulations and models that more accurately reflect the reality we are trying to recreate and often cast forward in time to figure out what will happen next.
Pushing performance is an interesting challenge here at the beginning of the 21st century, since a lot of server and storage components are commoditized and therefore available to others. The real engineering is coming up with innovative ways of putting …
DDN Turns The Crank On “Wolfcreek” Storage was written by Timothy Prickett Morgan at The Next Platform.
At some point, all of the big public cloud providers will have to eat their own dog food, as the parlance goes, and run their applications atop the cloudy version of their infrastructure that they sell to other people, not distinct and sometimes legacy systems that predate the ascent of their clouds. In this regard, none of the cloud providers are any different from any major enterprise or government agency that struggles with any kind of legacy system.
Search engine and online advertising giant Google wants its Cloud Platform business to compete against Amazon Web Services and Microsoft Azure and …
Google Wants Kubernetes To Rule The World was written by Timothy Prickett Morgan at The Next Platform.
In many ways, enterprises and hyperscalers have it easy. Very quickly in the wake of its announcement more than two decades ago, the Java programming language, a kind of virtualized C++, became the de facto standard for coding enterprise applications that run the business. And a slew of innovative storage and data analytics applications that have transformed computing were created by hyperscalers in Java and often open sourced so enterprises could use them.
The HPC community – and it is probably more accurate to say the many HPC communities – has it a bit tougher because they use a variety …
Chasing The Dream Of Code HPC Once, Run Anywhere was written by Timothy Prickett Morgan at The Next Platform.
In the worlds of high performance computing (HPC) and physics, seemingly straightforward challenges are frequently not what they seem at first glance.
For example, doing justice to an outwardly simple physics experiment involving a pendulum and drive motor can involve the need to process billions of data points. Moreover, even when aided by the latest high performance technology, such as the Intel Xeon Phi processor, achieving optimal compute levels requires ingenuity for addressing unexpected coding considerations.
Jeffery Dunham, the William R. Kenan Jr. Professor of Natural Sciences at Middlebury College in Vermont, should know. For about eight years, Professor Dunham …
Physics Code Modifications Push Xeon Phi Peak Performance was written by Nicole Hemsoth at The Next Platform.
Waiting for a simulation to complete before visualizing the results is often an unappealing prospect for researchers.
Verifying that output matches expectations early in a run helps prevent wasted computation time, which is particularly important on systems in high demand or when a limited allocation is availableIn addition, the growth in the ability to perform computation continues to outpace the growth in the ability to performantly store the results. The ability to analyze simulation output while it is still resident in memory, known as in situ processing, is appealing and sometimes necessary for researchers running large-scale simulations.
In light of …
Advances in In Situ Processing Tie to Exascale Targets was written by Nicole Hemsoth at The Next Platform.
It has been six years now since the “Austin” release of the OpenStack cloud controller was released by the partnership of Rackspace Hosting, which contributed its Swift object storage, and NASA, which contributed its Nova compute controller. NASA was frustrated by the open source Eucalyptus cloud controller, which was not completely open source and which did not add features fast enough, and Rackspace was in a fight for mindshare and marketshare against much larger cloud rival Amazon Web Services and wanted to leverage both open source and community to push back against its much larger rival.
OpenStack may not have …
Building The Stack Above And Below OpenStack was written by Timothy Prickett Morgan at The Next Platform.
Let’s be honest. Although the old saying “slow and steady wins the race” may be a lesson that helps us get through school, it isn’t a realistic credo for the unrelenting demands of today’s fast-paced businesses. Faster may be better, but only if quality doesn’t suffer – and this puts immense strain on agile organizations that must continually deliver new features and software to their customers.
Meeting these needs, and doing so with efficiency, requires rethinking how we view application development and operations. For organizations embracing and addressing these challenges, the pursuit of DevOps is the new normal, but it …
Getting Agile And Staying That Way was written by Timothy Prickett Morgan at The Next Platform.
Getting the ratio of compute to storage right is not something that is easy within a single server design. Some workloads are wrestling with either more bits of data or heavier file types (like video), and the amount of capacity required per given unit of compute is much higher than can fit in a standard 2U machine with either a dozen large 3.5-inch drives or two dozen 2.5-inch drives.
To attack these use cases, Cisco Systems is tweaking a storage-dense machine it debuted two years ago, and equipping it with some of the System Link virtualization technologies that it created …
Cisco Drives Density With S Series Storage Server was written by Timothy Prickett Morgan at The Next Platform.