Supercomputing Part 2

Supercomputing Part 2

This is a continuation of my previous post on Supercomputing 2022.

To say, the supercomputing space and the Supercomputing event (held in has changed would be a massive understatement.?

Like the performance of the systems that make up the Top 500 list, the event was so much (much) bigger. The conference’s scope too has expanded. Not only did this year’s attendees come from academia/research and HPC affiliated teams in the enterprise, but I saw quite a few attendees from adjacent groups. This includes teams responsible for AI/ML and data management from life sciences, manufacturing, and more. The Supercomputing conference has become the showcase for the broader HPC eco-system and the latest in AI/ML.

For example, WEKA partner Cerebras showcased their technology behind their Andromeda system delivering 1 Exaflop of AI compute and 120 Petaflops of dense compute. They spoke about their 13.5 million-core AI supercomputer (yes, we’re talking millions). This system delivers near perfect linear scaling for the largest GPT-class language models.?

So, why am I writing about this??

First, I am excited by the potential WEKA has to help organizations with not only their storage needs for traditional HPC workflows (ex: physics research, CAD/CAE, etc), but also the emerging AI/ML workflows (ex: dealing with PBs and PBs of data used to build and reinforce models, large scale interference workloads like fraud detection, and more). It was evident that the AI/ML teams in attendance were realizing the complexity of data related challenges they would be facing with machine learning and deep learning:

  1. They not only use traditional CPUs, but also rely on accelerated compute like GPUs and IPUs.?
  2. They deal with different types of data at a much different scale – terabytes to petabytes per project.
  3. They have to support a very diverse workflow – ingest and transform data, random read and write millions of small files in addition to sequential reads and writes, deal with metadata of thousands to millions of files, and more.
  4. They deal with different clouds – their workflows span different clouds. It’s not uncommon to burst from on-prem/private clouds to public clouds to handle a surge of learning activity.?

Second, to marvel on how this conference has evolved as the perfect? place for organizations to experience all parts of the ecosystem come together.? As someone responsible for building a vibrant solutions and ecosystem at WEKA, I saw almost all of our ecosystem partners in full force–from accelerated compute and networking like AMD, Cerebras, Habana & Intel, NVIDIA, and more to ISVs dealing with AI/ML, Data Management, and Workflow such as Ansys, Altair, Run:ai, and more to our strategic cloud and system partners like AWS, Dell, Google Cloud, HPE, Hitachi Vantara, Lenovo, Oracle, Supermicro, and more. An example of this in action is our partner Run:ai, who were connecting with and successfully educating organizations to talk about AI Computing and Machine Learning initiatives including areas like ML Operations (aka MLOps) and getting the most out of their GPU investments across different clouds.?

Third, because of the role our WEKA team played in educating attendees about the state of the art and future of Supercomputing – not only on storage, but across the entire stack. WEKA hosted dozens of partners at our booth theater at the conference, who spoke about industry trends, state of the art technologies, joint solutions, and real world examples – all intended to help organizations understand how they can tackle their next generation of application workflows.?

I’ll cover the learning I took from these sessions in my next post.

要查看或添加评论,请登录

Sachin Chheda的更多文章

  • Service Providers and End-User Computing: Desktop as a Service

    Service Providers and End-User Computing: Desktop as a Service

    This is joint post with Maryam Alexandrian-Adams (#Nutanix VDI lead). A lot has been said and done on the Virtual…

    3 条评论
  • Build Profitable and Differentiated AI Services With Nutanix

    Build Profitable and Differentiated AI Services With Nutanix

    As a technology investment, AI has been topping the list for most organizations. The 2025 Nutanix Enterprise Cloud…

    4 条评论
  • AI Technology Projects: The Case for New Skills

    AI Technology Projects: The Case for New Skills

    Source: TOI / Cookr's LI post There are a new crop of start ups looking at disrupting knowledge-intensive areas with a…

    1 条评论
  • Rise of Agentic AI

    Rise of Agentic AI

    I had a chance to attend an event early this month on the role of AI in the enterprise. Like it is with many of these…

    2 条评论
  • Serving Cloud-Native (and Other) Applications

    Serving Cloud-Native (and Other) Applications

    In a previous article I discussed considerations when ‘Thinking Cloud-Native’. I am switching gears in this article to…

    1 条评论
  • Sparking Creativity

    Sparking Creativity

    I recently posted on LinkedIn about critical thinking and reasoning. I talked about the need to be creative when…

    4 条评论
  • Building a Strong Foundation for Your Career

    Building a Strong Foundation for Your Career

    Recently I had a chance to talk to someone early in their career on a cross pacific flight. We ended up talking about…

    13 条评论
  • Thinking Cloud Native

    Thinking Cloud Native

    This is the first of many blogs of the topic of Cloud Native. This one covers the concept of Cloud Native and…

    2 条评论
  • Getting to Know Your (New) Hybrid Multicloud

    Getting to Know Your (New) Hybrid Multicloud

    A few years ago, I wrote a series of blogs on modernizing infrastructure and becoming cloud smart. The industry at that…

    3 条评论
  • Supercomputing Part 3

    Supercomputing Part 3

    In the previous two posts (here and here) I spoke about my own history with the Supercomputing space, evolution of the…

社区洞察

其他会员也浏览了