Archive

Category Archives for "The Next Platform"

Memory at the Core of New Deep Learning Research Chip

Over the last two years, there has been a push for novel architectures to feed the needs of machine learning and more specifically, deep neural networks.

We have covered the many architectural options for both the training and inference sides of that workload here at The Next Platform, and in doing so, started to notice an interesting trend. Some companies with custom ASICs targeted at that market seemed to be developing along a common thread—using memory as the target for processing.

Processing in memory (PIM) architectures are certainly nothing new, but because the relatively simple logic units inside of

Memory at the Core of New Deep Learning Research Chip was written by Nicole Hemsoth at The Next Platform.

Many Life Sciences Workloads, One Single System

The trend at the high end, from supercomputer simulations to large-scale genomics studies, is to push heterogeneity and software complexity while reducing the overhead on the infrastructure side. This might sound like a case of dueling forces, but there is progress in creating a unified framework to run multiple workloads simultaneously on one robust cluster.

To put this into context from a precision medicine angle, Dr. Michael McManus shared his insights about the years he spent designing infrastructure for life sciences companies and research. Those fields have changed dramatically in just the last five years alone in terms of data

Many Life Sciences Workloads, One Single System was written by Nicole Hemsoth at The Next Platform.

Orchestrating HPC Engineering in the Cloud

Public clouds have proven useful to a growing number of organizations looking for ways to run their high-performance computing applications to scale without having to limit themselves to whatever computing capabilities they have in-house or to spending a lot of money to build up their infrastructure to meeting their growing needs.

The big three – Amazon Web Services, Microsoft Azure and Google Cloud – have rolled out a broad array of compute, networking and storage technologies that companies can leverage when their HPC workloads scale to the point that they can no longer be run on their in-house workstations or

Orchestrating HPC Engineering in the Cloud was written by Nicole Hemsoth at The Next Platform.

ARM Server Chips Challenge X86 in the Cloud

The idea of ARM processors being used in datacenter servers has been kicking around more most of the decade. The low-power architecture dominates the mobile world of smartphones and tablets as well as embedded IoT devices, and with datacenters increasingly consuming more power and generating more heat, the idea of using highly efficient ARM chips in IT infrastructure systems gained steam.

That was furthered by the rise of cloud computing environments and hyperscale datacenters, which can be packed with tens of thousands of small servers running massive numbers of workloads. The thought of using ARM-based server chips that are more

ARM Server Chips Challenge X86 in the Cloud was written by Nicole Hemsoth at The Next Platform.

Riding The Coattails Of Google Kubernetes And AWS Lambda

There are individuals and companies that create whole new technologies for their own consumption and that sometimes open source them for others to help steer their development and fix their bugs. And then there are still other companies that polish these tools, giving them some enterprise fit and finish, and thereby make it possible for others to deploy a particular technology without having to have PhDs, who are not available anyway, on staff.

From the enterprise perspective, the Apache web server and related Tomcat application server needed its Big Blue, the Linux operating system needed its Red Hat, and the

Riding The Coattails Of Google Kubernetes And AWS Lambda was written by Timothy Prickett Morgan at The Next Platform.

Veteran IT Journalist, Jeffrey Burt, Joins The Next Platform as Senior Editor

We are thrilled to announce the full-time addition of veteran IT journalist, Jeffrey Burt to The Next Platform ranks.

Jeffrey Burt has been a journalist for more than 30 years, with the last 16-plus year writing about the IT industry. During his long tenure with eWeek, he covered a broad range of subjects, from processors and IT infrastructure to collaboration, PCs, AI and autonomous vehicles.

He’s written about FPGAs, supercomputers, hyperconverged infrastructure and SDN, cloud computing, deep learning and exascale computing. Regular readers here will recognize that his expertise in these areas fits in directly with our coverage

Veteran IT Journalist, Jeffrey Burt, Joins The Next Platform as Senior Editor was written by Nicole Hemsoth at The Next Platform.

OpenCL Opens Doors to Deep Learning Training on FPGA

Hardware and device makers are in a mad dash to create or acquire the perfect chip for performing deep learning training and inference. While we have yet to see anything that can handle both parts of the workload on a single chip with spectacular results (the Pascal general GPUs are the closest thing yet, with threats coming from Intel/Nervana in the future), there is promise for FPGAs to find inroads.

So far, most of the work we have focused on for FPGAs and deep learning has centered more on the acceleration of inference versus boosting training times and accuracy

OpenCL Opens Doors to Deep Learning Training on FPGA was written by Nicole Hemsoth at The Next Platform.

IARPA Spurs Race to Speed Cryogenic Computing Reality

The race is on to carve a path to efficient extreme-scale machines in the next five years but existing processing approaches fall far short of the efficiency and performance targets required. As we reported at the end of 2016, the Department of Energy in the U.S. is keeping its eye on non-standard processing approaches for one of its exascale-class systems by 2021, and other groups, including the IEEE are equally keeping pace with new architectures to explore as CMOS alternatives.

While there is no silver bullet technology yet that we expect will sweep current computing norms, superconducting circuits appear

IARPA Spurs Race to Speed Cryogenic Computing Reality was written by Nicole Hemsoth at The Next Platform.

Hyperscalers Ready To Run Barefoot In The Datacenter

Breaking into the switch market is not an easy task, whether you are talking about providing whole switches or just the chips that drive them. But there is always room for innovation, which is why some of the upstarts have a pretty credible chance to shake up networking, which is the last bastion of proprietary within the datacenter.

Barefoot Networks is one of the up-and-coming switch chip makers, with its “Tofino” family of ASICs that, among other things, has circuits and software that allow for the data plane – that part of the device that controls how data moves

Hyperscalers Ready To Run Barefoot In The Datacenter was written by Timothy Prickett Morgan at The Next Platform.

Looking Through the Windows at HPC OS Trends

High performance computing (HPC) is traditionally considered the domain of large, purpose built machines running some *nix operating system (predominantly Linux in recent years). Windows is given little, if any, consideration. Indeed, it has never accounted for even a full percent of the Top500 list. Some of this may be due to technical considerations: Linux can be custom built for optimum performance, including recompiling the kernel. It is also historically more amenable to headless administration, which is a critical factor when maintaining thousands of nodes.

But at some point does the “Windows isn’t for high-performance computing” narrative become self-fulfilling?

Looking Through the Windows at HPC OS Trends was written by Nicole Hemsoth at The Next Platform.

The Relentless Yet Predictable Pace Of InfiniBand Speed Bumps

High performance computing in its various guises is not just determined by the kind and amount of computing that is made available at scale to applications. More and more, the choice of network adapters and switches as well as the software stack that links the network to applications plays an increasingly important role. And moreover, networks are comprising a larger and larger portion of the cluster budget, too.

So picking the network that lashes servers to each other and to their shared storage is important. And equally important is having a roadmap for the technology that is going to provide

The Relentless Yet Predictable Pace Of InfiniBand Speed Bumps was written by Timothy Prickett Morgan at The Next Platform.

Skylake Xeon Ramp Cuts Into Intel’s Datacenter Profits

Every successive processor generation presents its own challenges to all chip makers, and the ramp of 14 nanometer processes that will be used in the future “Skylake” Xeon processors, due in the second half of this year, cut into the operating profits of its Data Center Group in the final quarter of 2016. Intel also apparently had an issue with one of its chip lines ­– it did not say if it was a Xeon or Xeon Phi, or detail what that issue was – that needed to be fixed and that hurt Data Center Group’s middle line, too.

Still,

Skylake Xeon Ramp Cuts Into Intel’s Datacenter Profits was written by Timothy Prickett Morgan at The Next Platform.

A Case for CPU-Only Approaches to HPC, Analytics, Machine Learning

With the current data science boom, many companies and organizations are stepping outside of their traditional business models to scope work that applies rigorous quantitative methodology and machine learning – areas of analysis previously in the realm of HPC organizations.

Dr. Franz Kiraly an inaugural Faculty Fellow at the Alan Turing Institute observed at the recent Intel HPC developer conference that companies are not necessarily struggling with “big” data, but rather with data management issues as they begin to systematically and electronically collect specific data in one place that makes analytics feasible. These companies, as newcomers to “machine learning” and

A Case for CPU-Only Approaches to HPC, Analytics, Machine Learning was written by Nicole Hemsoth at The Next Platform.

AWS Outlines Current HPC Cloud User Trends

Last week we discussed the projected momentum for FPGAs in the cloud with Deepak Singh, general manager of container and HPC projects at Amazon Web Services. In the second half of our interview, we delved into the current state of high performance computing on Amazon’s cloud.

While the company tends to offer generalizations versus specific breakdowns of “typical” workloads for different HPC application types, the insight reveals a continued emphasis on pushing new instances to feed both HPC and machine learning, continued drive to push ISVs to expand license models, and continued work to make running complex workflows more seamless.

AWS Outlines Current HPC Cloud User Trends was written by Nicole Hemsoth at The Next Platform.

FPGAs Focal Point for Efficient Neural Network Inference

Over the last couple of years, we have focused extensively on the hardware required for training deep neural networks and other machine learning algorithms. Focal points have included the use of general purpose and specialized CPUs, GPUs, custom ASICs, and more recently, FPGAs.

As the battle to match the correct hardware devices for these training workloads continues, another has flared up on the deep learning inference side. Training neural networks has its own challenges that can be met with accelerators, but for inference, the efficiency, performance, and accuracy need to be in balance.

One developing area in inference is in

FPGAs Focal Point for Efficient Neural Network Inference was written by Nicole Hemsoth at The Next Platform.

Heating Up the Exascale Race by Staying Cool

High performance computing is a hot field, and not just in the sense that it gets a lot of attention. The hardware necessary to perform the countless simulations performed every day consumes a lot of power, which is largely turned into heat. How to handle all of that heat is a subject that is always on the mind of facilities managers. If the thermal energy is not moved elsewhere in short order, the delicate electronics that comprise the modern computer will cease to function.

The computer room air handler (CRAH) is the usual approach. Chilled water chills the air, which

Heating Up the Exascale Race by Staying Cool was written by Nicole Hemsoth at The Next Platform.

IBM Reorg Forges Cognitive Systems, Merges Cloud And Analytics

It is the first month of a new year, and this is the time that IBM traditionally does reorganizations of its business lines and plays musical chairs with its executives to reconfigure itself for the coming year. And just like clockwork, late last week the top brass at Big Blue did internal announcements explaining the changes it is making to transform its wares into a platform better suited to the times.

The first big change, and one that may have precipitated all of the others that have been set in place, is Robert LeBlanc, who is the senior vice president

IBM Reorg Forges Cognitive Systems, Merges Cloud And Analytics was written by Timothy Prickett Morgan at The Next Platform.

Refining Oil and Gas Discovery with Deep Learning

Over the last two years, we have highlighted deep learning use cases in enterprise areas including genomics, large-scale business analytics, and beyond, but there are still many market areas that are still building a profile for where such approaches fit into existing workflows. Even though model training and inference might be useful, for some areas that have complex simulation-driven workflows, there are great efficiencies that could come from deep neural nets, but integrating those elements is difficult.

The oil and gas industry is one area where deep learning holds promise, at least in theory. For some steps in the resource

Refining Oil and Gas Discovery with Deep Learning was written by Nicole Hemsoth at The Next Platform.

Multi-Threaded Programming By Hand Versus OpenMP

For a long time now, researchers have been working on automating the process of breaking up otherwise single-threaded code to run on multiple processors by way of multiple threads. Results, although occasionally successful, eluded anything approaching a unified theory of everything.

Still, there appears to be some interesting success via OpenMP. The good thing about OpenMP is that its developers realized that what is really necessary is for the C or Fortran programmer to provide just enough hints to the compiler that say “Hey, this otherwise single-threaded loop, this sequence of code, might benefit from being split amongst multiple

Multi-Threaded Programming By Hand Versus OpenMP was written by Timothy Prickett Morgan at The Next Platform.

AWS Details FPGA Rationale and Market Trajectory

At the end of 2016, Amazon Web Services announced it would be making high-end Xilinx FPGAs available via a cloud delivery model, beginning first in a developer preview mode before branching with higher-level tools to help potential new users onboard and experiment with FPGA acceleration as the year rolls on.

As Deepak Singh, General Manager for the Container and HPC division within AWS tells The Next Platform, the application areas where the most growth is expected for cloud-based FPGAs are many of the same we detailed in our recent book, FPGA Frontiers: New Applications in Reconfigurable Computing. These

AWS Details FPGA Rationale and Market Trajectory was written by Nicole Hemsoth at The Next Platform.