Our website uses cookies to give you the most optimal experience online by: measuring our audience, understanding how our webpages are viewed and improving consequently the way our website works, providing you with relevant and personalized marketing content.
You have full control over what you want to activate. You can accept the cookies by clicking on the “Accept all cookies” button or customize your choices by selecting the cookies you want to activate. You can also decline all non-necessary cookies by clicking on the “Decline all cookies” button. Please find more information on our use of cookies and how to withdraw at any time your consent on our privacy policy.

Managing your cookies

Our website uses cookies. You have full control over what you want to activate. You can accept the cookies by clicking on the “Accept all cookies” button or customize your choices by selecting the cookies you want to activate. You can also decline all non-necessary cookies by clicking on the “Decline all cookies” button.

Necessary cookies

These are essential for the user navigation and allow to give access to certain functionalities such as secured zones accesses. Without these cookies, it won’t be possible to provide the service.
Matomo on premise

Marketing cookies

These cookies are used to deliver advertisements more relevant for you, limit the number of times you see an advertisement; help measure the effectiveness of the advertising campaign; and understand people’s behavior after they view an advertisement.
Adobe Privacy policy | Marketo Privacy Policy | Pardot Privacy Policy | Oktopost Privacy Policy | MRP Privacy Policy | AccountInsight Privacy Policy | Triblio Privacy Policy

Social media cookies

These cookies are used to measure the effectiveness of social media campaigns.
LinkedIn Policy

Our website uses cookies to give you the most optimal experience online by: measuring our audience, understanding how our webpages are viewed and improving consequently the way our website works, providing you with relevant and personalized marketing content. You can also decline all non-necessary cookies by clicking on the “Decline all cookies” button. Please find more information on our use of cookies and how to withdraw at any time your consent on our privacy policy.

Skip to main content

The three variations of supercomputing cloud technologies

“Cloud computing” is one of the most overloaded terms in the history of information technology. Yes, there is the NIST Definition of Cloud Computing that many will point to, but the sheer number of adaptations of models for “enabling ubiquitous, convenient, on-demand” access to computing resources knows no limit. When we narrow the aperture a bit, even to broad horizontal use cases such as supercomputing and HPC, things become a bit clearer. Partly, this has to do with the relatively late adoption of cloud-like concepts in this space.

I remember having conversations in 2013 and 2014 with “birds of a feather” who simply laughed at the idea of doing “proper” HPC in the cloud. Cloud generally meant cheap storage and compute – the antithesis of supercomputing. Obviously, things are different today, but it’s not simply because of the technological advances in the space. The industry managed to take a step back from terminology and purity and think about actual unsolved problems, and how new paradigms could help if adapted the right way.

For the Atos Nimbix Supercomputing Suite, we applied this logic to three major and very different types of issues and found that a common technology thread worked across the board. In fact, we also learned that while the models are different, organizations often benefit from more than one, and in some cases all three – whether immediately or over time.

Elastic Supercomputing

The most “cloud-like” of the three models, elastic caters to researchers and engineers who are looking for on-demand access to advanced computing and applications without complex procurement, configuration, build-out, and tuning processes. They simply log into a web portal, choose a workflow and a dataset, and run it. Most elastic users can get started in minutes. This is software-as-a-service (SaaS) for HPC.

HPC, AI, Quantum supercomputing in minutes

LOG IN
to your account

SELECT
your application

UPLOAD
your files

CHOOSE
machine type and cores

RUN
your workloads

Over the years, we found the most obvious inflection point to be when workstations are simply not enough to run solvers at the required scale and performance. Organizations that don’t already have clusters often find it quite difficult and cost-prohibitive to level up to large-scale HPC. It means buying equipment, leasing datacenter space, hiring experts, and spending months putting everything together. There are plenty of consultants and services out there to help, but obviously at a commensurate cost. For the engineers needing to get work done faster, the appeal of simply running the same application they already do on their workstation but at 8X or 16X the scale for a fraction of the price is hard to ignore. There are plenty of cost models that would argue otherwise, but all require major investment up front in both time and resource, and none solve the problem “right now” like SaaS does.

What’s more, the elastic model even helps those organizations that already have HPC in-house. Whether it’s expanding access for users who are not as skilled in writing traditional job scheduler scripts (but are otherwise brilliant in their respective science/engineering domain), or quickly augmenting capacity during peak periods without lengthy up-front setup or commitment, HPC SaaS adds practical options at a fair price. The Nimbix Elastic offering even supports federated multi-zone usage to respect data sovereignty and governance across geographies. In short, it can help any organization, even a multi-national one, get more work done quickly and with virtually no up-front effort.

Dedicated Supercomputing

Physical, on-premises (or even privately hosted) infrastructure is generally at odds with the concepts of cloud computing, but its procurement and consumption models need not be. It’s true that many organizations have specific requirements that only dedicated clusters can meet. Cloud enthusiasts will argue that the largest, most powerful compute clusters in the world are already available on demand from various “hyperscale” companies, but particularly with HPC, real-world offerings may still not meet all customer requirements. And of course, there is cost – full-time use of systems gets expensive when paying on-demand prices.

But customers needing control of their own infrastructure still face two problems that cloud-like solutions, whether the technological kind or the contractual kind, can easily address.

First, there’s the need for OPEX-style financing, particularly of what is naturally a depreciating asset. Additionally, considering most systems may not be utilized to their full capabilities at all times, tying up capital is a difficult proposition to make. Paying for hardware “as you go”, as if it were infrastructure-as-a-service, can be attractive, especially if you factor in growth and elasticity needs. When customers need more hardware, they simply pay a larger monthly bill and have immediate access to it rather than entering an entire new procurement cycle to purchase all of it at once. This is not unlike leasing, except with more elasticity and flexibility.

The second problem is about usability and access – again, similar in scope to what drives some customers to the SaaS model. Continuing to buy systems that only experts in both domain science and information technology can use limits innovation, particularly as demand for talent becomes more and more specialized. Why preclude the best engineers and researchers in their respective fields simply because they are not highly skilled at interacting with scheduler software developed decades ago for batch computing purposes? It doesn’t make a lot of sense.

Democratization of HPC helps everyone, regardless of how compute capacity is delivered.
Nimbix Dedicated was designed to solve both problems, with an on-demand financial model for procuring equipment, and seamlessly enabling our HPC-as-a-service application delivery platform, JARVICE™ XE. What’s more, bursting to other infrastructure, including public clouds, is a simple

matter of pointing and clicking in the modern web-based user interface. While this may not be an immediate need for some organizations, it’s in almost everyone’s plans for “next generation” supercomputing.

Federated Supercomputing

The age-old problem of what to do with spare computing capacity is no different in HPC as it’s been for decades in the mainframe world. Customers pay for lots of compute, and they don’t always use it. Rather than relying on a vendor to help drive utilization, why not offer this capacity to the market in exchange for fees to help fund its ongoing operation and financing?

With Nimbix Federated, we addressed this challenge in two ways. The first is a managed global control plane that any organization can join simply by deploying software components and connecting to it – in essence becoming service providers but with minimal operational overhead. These newly minted providers decide what resources to make available, at what cost, and to what customer or customers. Atos does the rest, operating the JARVICE XE platform to securely manage all the infrastructure, including billing and accounting. When customers use a provider’s cycles, that provider gets paid.

The second is with a “private” federation, where a group of providers join forces to exchange capabilities, capacity, and even encourage collaboration. Rather than attaching to a public control plane, they deploy and operate their own. Atos can manage it for them as well if they like. This way, a related group of institutions can make their own rules, ensuring policies and governance are always met to their precise standards (e.g., when dealing with sensitive data such as healthcare, energy, and defense). But rather than each provider managing all users and settings for all “client” organizations, they simply connect to the control plane and let it handle it for them. Once again, the JARVICE XE platform facilitates all this.

When three very different (but often converging) paradigms rely on similar technology, everyone wins. Transitioning between delivery models becomes basically transparent to end users, ensuring innovation above all else continues at the blinding pace the world needs. This is, after all, supercomputing’s raison d’etre.

Share this blog article


About Leo Reiter
CTO and Technical Director of the Atos HPC Cloud Competency Center


Leo has spent his entire career in large-scale and server-based computing, in virtually every role from hands-on software architect to senior executive. His organizational experience ranges from early-stage to Global 2000/Fortune 500, and everything in between. He has led and/or played key roles in creating and launching products and lines of business, evangelism and brand awareness, acquiring patents, recruiting and retaining talent, fundraising, organizational change, pivots and go-to-market strategy, strategic partnerships, and corporate development/M&A process. At Nimbix, he built the software, intellectual property, product team, and problem-solving culture that transformed the company from a boutique internet services provider to a global platform software and SaaS leader in the High-Performance Computing (HPC) space, culminating in a successful acquisition by Atos in 2021. Prior to Nimbix he co-founded and served as Chief Technology Officer (CTO) of Virtual Bridges, the company that invented Virtual Desktop Infrastructure and pioneered ephemeral workloads delivered from data centers and centrally-managed edge devices – better known today as hybrid cloud computing. Earlier in his career, he worked as a software developer, technical leader, and consultant in various high-tech sectors such as semiconductors, industrial automation, and telecommunications.