HPE wants to turn supercomputing leadership into generative AI profits
Hewlett Packard Enterprise Co.’s recent announcement of an artificial intelligence cloud for large language models highlights a differentiated strategy that the company hopes will lead to sustained momentum in its high-performance computing business.
Although we think HPE has some distinct advantages with respect to its supercomputing intellectual property, the public cloud players have a substantial lead in AI with a point of view that generative AI such as OpenAI LP’s ChatGPT is fully dependent on the cloud and its massive compute capabilities. The question is: Can HPE bring unique capabilities and a focus to the table that will yield competitive advantage and ultimately, profits in the space?
In this Breaking Analysis, we unpack HPE’s LLM-as-a-service announcements from the company’s recent Discover conference and we’ll try to answer the question: Is HPE’s strategy a viable alternative to today’s public and private cloud gen AI deployment models, or is it ultimately destined to be a niche player in the market? To do so, we welcome to the program CUBE analyst Rob Strechay and Andy Thurai, vice president and principal analyst from Constellation Research Inc.
HPE announces an AI cloud
In 2014, prior to the split of HP and HPE, HP announced the Helion public cloud. Two years later it shut down the project and ceded the public cloud to Amazon Web Services Inc. At the time, HPE lacked the scale and differentiation to compete.
The company hopes this time around will be different. This past week at its Discover event, HPE entered the AI cloud market via an expansion of its GreenLake as-a-service platform. The company is offering large language models on-demand in a multitenant service, powered by HPE supercomputers.
HPE is partnering with a Germany-based startup called Aleph Alpha GmbH, a company specializing in large language models with a focus on explainability. HPE believes that this is critically important for its strategy of offering domain-specific AI applications. HPE’s first offering will provide access to Luminous, a pretrained LLM from Aleph Alpha what will allow companies to leverage their own data to train and tune custom models using proprietary information.
We asked Strechay and Thurai to unpack the announcement and provide their perspective. Here’s a summary of that conversation:
The core of the discussion centers on HPE’s plans to utilize Cray supercomputing infrastructure in an “as-a-service” model, aiming to make high-performance computing more accessible.
The following key points are noteworthy:
- Strechay acknowledges HPE’s innovative approach of providing supercomputing power as a service, leveraging its Cray technology, but points out the announcement precedes the actual general availability by about six months. He suggests that HPE is playing catch-up in the LLM market but acknowledges it’s coming at it from a novel angle of high-performance computing.
- Thurai agrees with Strechay’s assessment but adds some optimism, suggesting that the proposed model could be compelling for large workloads. He finds the idea of users being able to pass their biggest workloads to HPE without needing to fine-tune anything to be compelling, particularly for high-performance computing tasks.
- However, Thurai also has concerns. He calls out the absence of concrete details about key aspects such as machine learning operations or MLOps in HPE’s announcement, and he emphasizes the need to see these before forming a solid opinion on the viability of HPE’s strategy.
- Strechay also points out that this offering should be thought of more as platform as a service vs. infrastructure as a service.
Bottom Line:
The analysts are cautiously optimistic about HPE’s announced strategy, noting that it could potentially revolutionize how large workloads and high-performance computing tasks are handled. However, both agree that the company needs to provide more specifics about its execution plan, particularly around MLOps, before any substantial conclusions can be drawn. Ultimately, it’s a matter of execution.
Narrow workload scope to sharpen focus
The conversation between Strechay and Thurai further delves into the specific workloads HPE plans to address with its new LLM-as-a-service offering, including climate modeling, bio-life sciences, healthcare and potentially financial modeling. The analysts also discuss HPE’s partnership with a lesser-known company, Aleph Alpha.
The following key points are noteworthy:
- Strechay identifies the three major sectors HPE plans to target – climate, bio-life sciences and healthcare. They suggest these sectors are those in which the Cray supercomputing infrastructure excels, and HPE’s approach to making this infrastructure available as-a-service can simplify the process for users.
- Strechay highlights the PaaS nature of the offering, emphasizing that users can either leverage HPE’s models or import their own, like those from Anthropic. This PaaS nature means users won’t have to go through the rigors of setting up their own infrastructure.
- Thurai discusses HPE’s partnership with Aleph Alpha, noting that the ultimate goal for HPE is to demonstrate its capability to handle the training of LLMs, which have become as demanding as traditional high-performance computing tasks.
- Thurai notes that he appreciates HPE’s demonstration but voices concerns about the lack of details on how HPE plans to handle diverse AI, machine learning and deep learning workloads. He also notes that though HPE has affinity toward open-source, the broader ecosystem of components required for a robust AI/ML service is still unclear.
Bottom line:
HPE’s new strategy offers promise in making supercomputing as-a-service a reality for significant sectors such as climate, healthcare and bio-life sciences. Its partnership with Aleph Alpha, though not with a mainstream company, signals a move toward demonstrating prowess in handling large AI workloads. While the direction seems promising, we still raise concerns about the absence of details around handling diverse AI, machine learning and deep learning workloads and the overall ecosystem approach.
I think in Europe, sustainability will significantly help them. I don’t think it’s as big an advantage in North America – Rob Strechay
Watch this clip of the analysts’ discussion, unpacking HPE’s AI Cloud and the prospects for success.
High-performance computing meets AI
HPE’s fundamental belief is that the worlds of high-performance computing and AI are colliding in a way that confers competitive advantage to HPE. Indeed, HPE has a leadership position in HPC, as shown below.
HPE is No. 1 and No. 3 in terms of the world’s top five supercomputers with its Frontier and Lumi systems. Both leverage HPE’s Slingshot interconnect, which it believes is a critical differentiator.
It also believes that generative AI’s unique workload characteristics favor HPE’s supercomputing expertise. Here’s how Dr. Eng Lim Goh, HPE’s chief technology officer for AI, describes the difference between traditional cloud workloads and gen AI:
The traditional cloud service model is where you have many, many workloads running on many computer servers. But with a large language model, you have one workload running on many computer servers. And therefore, the scalability part is very different. This is where we bring in our supercomputing knowledge that we have for decades to be able to deal with this one big workload on many computer servers.
Here’s a summary of the analysts’ discussion:
Strechay and Thurai dive deeper into HPE’s legacy and potential within the large language model market, while analyzing the challenges they may face. Strechay draws on the company’s rich history in handling large applications, suggesting that this experience could give them a certain advantage. Thurai, however, seems skeptical about the company’s ability to leverage these resources and align them with the market’s needs.
The following additional points are noteworthy:
- Strechay acknowledges HPE’s history and pedigree in managing large applications across many servers, drawing upon its involvement in the Open Grid Forum and Global Grid Forum. He sees HPE’s longstanding relationships and experiences with significant entities such as NASA and the Department of Energy as a potential competitive advantage.
- Thurai questions the mainstream appeal of HPE’s service, suggesting that it’s more niche-oriented. He disagrees with the suggestion that HPE is the only one offering HPC services, mentioning Amazon’s HPC service as an example of a robust competitor.
- Thurai raises concerns about the amount of data accessible to HPE relative to the public cloud, pointing out that much of the innovation and AI workloads will go to the hyperscalers thanks to data accessibility. He acknowledges HPE’s powerful supercomputer and storage capacity but questions whether its promise to handle the most extensive workloads will be sufficient to move the needle in its favor.
Bottom line:
HPE’s vast experience and pedigree in managing extensive applications and longstanding relationships might provide it an advantage in the large language model market. However, potential challenges in data access for innovation workloads and the competitiveness of the market may pose hurdles to HPE’s success. Although the company boasts a powerful supercomputer and storage capacity, its ability to turn these assets into a compelling offering that outperforms rivals remains uncertain.
Follow the money: breaking down HPE’s business segments
Above, we take a look at HPE’s lines of business and how its AI and HPC line of business perform. Remember HPE purchased Cray Inc. in 2019 and Silicon Graphics Inc. a few years before that to get into the HPC market.
Looking at HPE’s most recent quarter, you can see how it reports its business segments. HPC and AI is a multibillion-dollar business – and it’s growing – but it essentially is a break-even business. So it brings bragging rights but not profits. Intelligent Edge – aka Aruba — is the shining star right now with a $5 billion-plus run rate and 27% operating profit – so marginwise it’s HPE’s best business and throws off nearly as much profit as its server business.
Here’s how HPE Chief Executive Antonio Neri describes its advantage:
If you think about how public clouds are being architected, it’s a traditional network architecture at massive scale, with leaf and spine where generic or general-purpose workloads of sorts use that architecture to run workloads and connect to the data. When you go to this [LLM] architecture, which is an AI-native architecture, the network is completely different. You mentioned Slingshot. That network runs and operates totally different. Obviously, you need the network interface cards that connect with each GPU or CPU. And also, a bunch of accelerators that come with it. And there is silicon programmability with the contention software management. And that’s what Slingshot is all about, and it takes many, many years to develop. But if you look at public clouds today, generally speaking, they have not developed a network. They have been using companies like Arista, Cisco or Juniper and the like. We have that proprietary network. And so does Nvidia, by the way. But ours actually opens up multiple ecosystems and we can support any of them. So, it will take a lot of time and effort [for clouds to catch up]. And then, also remember, you’re now dealing with a whole different compute stack, which is direct liquid cooling, and that requires a whole different set of understanding.
There’s a lot to unpack in terms of what Antonio stated, including the network, Slingshot interconnect, the data services ecosystem and liquid cooling. We asked the question: “Is this a flip on ‘Jassy’s Law’ – i.e. there’s no compression algorithm for experience? Or does HPE have blind spots?”
The following points summarize the analysts’ take:
- Thurai suggests that though HPE may seem to be highly involved in AI, most of its work is still in classic HPC workloads. Its focus on AI workloads, including LLM, appears more as a demonstration rather than actual work with these types of systems. Thurai expresses skepticism about HPE’s ability to persuade users to run LLM workloads on their servers because of a lack of an ecosystem and an MLOps presence.
- Thurai further notes that to train models, partnerships are necessary, such as with a company such as Hugging Face, something HPE did not put forth. Contrasting this with AWS’ approach, HPE’s strategy seems to be focused on rebranding classic HPC workloads as AI workloads, a move whose success reasonable folks will question.
Bottom line:
The key issue is whether HPE’s strategy of focusing on classic HPC workloads can be profitable. Although HPE’s network and interconnect give the company potential advantages, these may be short-lived, as commercial components are available off-the-shelf. Expertise with liquid cooling in data centers is nice but the real questions will come down to HPE’s ability to attract customer data to its platform versus those of competitors.
How IT decision makers are thinking about using gen AI and LLMs
The next question we want to explore is: Does HPE’s service have the potential to go mainstream or is it destined for niche status?
Above we show some Enterprise Technology Research data asking organizations how they’re pursuing generative AI and LLMs and what use cases they’re evaluating or actively deploying in production. Note that 34% of the organizations say they’re not evaluating, which is surprisingly high in our view. But for those moving forward, the use cases are what you’d expect, chatbots, code generation, writing marketing copy, summarizing text and more.
HPE has a different point of view. It’s focusing on very specific domains where companies have their own proprietary data and want to train that data but don’t want to incur the expense of acquiring and managing their own supercomputing infrastructure. At the same time, HPE believes because it has unique IP, it can be more reliable and cost-effective than the public cloud players, while still offering the advantages of a public cloud.
We asked the analysts: Is HPE on to something here in that these mainstream use cases are not where the money is for HPE? And is there gold in the hills with HPE’s strategy?
Although generally we’re taking a wait-and-see, “show me” approach with HPE’s LLM strategy, the following points are notable:
- We do believe HPE can find a profitable niche by providing supercomputing capabilities to those that don’t have the means or resources to invest heavily in this area. It does not necessarily need to compete directly with major players like Amazon.
- That said, Thurai distinguishes between innovation workloads and mature workloads in the AI domain. For innovation workloads, priorities revolve around experimentation and speed. Sustainability, carbon footprint and cost-efficiency are important but will be overlooked in his view. Priorities will shift when the AI model matures, and issues such as security, governance, ethics, explainability, sustainability and liability become more important.
- HPE is positioning itself as the go-to solution for these mature workloads, taking on the complex tasks associated with them. If it can communicate this effectively, and the market aligns with its strategy, it could see success, and sustainability could become a significant factor. Furthermore, this strategy does differentiate HPE and comes at the problem from its position of HPC strength.
Bottom line:
HPE’s strategy caters to a specific sector of the AI market – those dealing with HPC workloads. This niche could offer profitable opportunities, given the specialized needs and complexities involved. However, their success relies on effectively communicating their value proposition and the alignment of market trends.
AWS: ‘Everything’s going to the cloud’ … HPE: ‘Uh – no, it’s not’
At Discover on the main stage we heard two distinct points of view:
Matt Wood of AWS was on the main stage with Neri, and much to our surprise, Wood said something to the effect of “over time, we still believe most of the workloads are going to go to the public cloud.” He actually said that in front of HPE’s audience.
Then, Neri basically countered that with (and we’re paraphrasing with tongue in cheek), “The world’s hybrid, dude. And it’s going to stay that way.”
Remember that scene in “Bridesmaids?” where the two bridesmaids are duel-singing for attention of the bride? Well, we heard this divergent theme with respect to LLMs this week. HPE put forth the notion that supercomputing workloads are different from cloud workloads and HPE has the expertise to make it happen more reliably, sustainable and effectively. Then on Bloomberg, we heard AWS CEO Adam Selipsky put forth the premise that LLMs are fully dependent on the public cloud and its massive compute capabilities.
At the end of “Bridesmaids,” the two rivals became good friends, so perhaps there’s room for both points of view. Although we believe the market in the public cloud for LLMs will be meaningfully larger, we don’t currently have a good enough sense of the delta to put a figure on it.
Here’s a summary of the analyst conversation:
- HPE needs to capitalize on its strengths, one of which is supercomputing. It may not directly compete with the public cloud, but it could carve out a unique space in the market.
- The next six months, leading up to general availability of HPE’s LLM cloud, will not be a singularly definitive period for HPE’s success in this market. Instead, we believe the process will have a “long tail,” indicating that the full impact and success of HPE’s strategy will unfold over a more extended period.
- Once models are trained and ready for production, considerations of sustainability (scope 1, 2, 3 emissions) and other factors become more important. The cloud could serve as a good “playground” for initial development and experimentation and models brought into HPE’s environment for production work.
Bottom line:
We believe HPE’s focus of leveraging its strengths, particularly in supercomputing, is sound. However, achieving success in its chosen AI market niche will likely be a long-term process and will require greater recognition from customers that HPE is a player in AI. To do so, the company will have to leverage its distribution channel to attract key partners that are known for their AI prowess.
HPE must cultivate AI mindshare through key partnerships
Despite extensive use of AI in its portfolio of offerings, HPE is not known as a player in AI. Let’s take a look at where the ETR data shows which firms are getting the share of wallet in the ML/AI market. Importantly, HPE has an opportunity to partner and accelerate its mindshare among tech decision makers.
The chart above shows Net Score or spending momentum on the vertical axis and the pervasiveness or presence in the ETR data set for ML/AI players. Right off the bat, focus on the Big Three public cloud players, Microsoft Corp., AWS and Google LLC – they dominate the conversation. They are pervasive and all show above the magic 40% red dotted line, an indicator of strong momentum.
Databricks Inc. also stands out as clearly a player in the mix.
OpenAI is also notable. We got a peek at the July ETR data and it won’t surprise you that OpenAI is setting new records — beyond even where we saw Snowflake Inc. at its peak Net Score. And OpenAI, as you’ll see next month in the ETR data, has gone mainstream, even in core IT shops.
It’s no surprise that you don’t see HPE in this mix, but over time, if the company’s aspirations are to come true, like Oracle Corp. and IBM Corp. you would want to see HPE on this chart.
Here are the key points from the analyst discussion:
- Many companies in the AI market are focusing on training large language models, retraining existing models and fine-tuning models.
- HPE, however, has taken a different approach. It is positioning itself to handle the most substantial and complex models, offering its clients the ability to leverage HPE’s strengths in computing, networking and storage.
- Questions remain as to whether HPE will succeed with this strategy. It will take at least another year to gauge the effectiveness of HPE’s approach and determine customer reaction.
Bottom line:
HPE’s strategy in the AI market differs from many competitors by focusing on handling the largest and most complex models, leveraging its high-performance computing, networking and storage strengths. The effectiveness of this approach remains to be seen and will likely take another year or so to evaluate.
One indicator to watch is how integrated HPE’s solution really is in the GreenLake console. Is it a separate console? Is it really on top of the Aruba Central platform or is it a separate installation?
Factors to watch
We close by discussing the competitive advantages and challenges HPE faces with some critical areas we’re watching.
Here’s a summary of our wrap up:
- We believe HPE’s competitive advantage may lie in the infrastructure software within its Cray-based technology, rather than just LLMs.
- HPE’s AI ecosystem is currently weak, lacking model repositories, model sharing and a robust software stack. HPE would likely need to partner with model producers to enhance this.
- A potential advantage for HPE could be the “bring your own model” approach, providing clients a hassle-free environment for training AI models with comprehensive support.
- Another potential advantage could be HPE’s ability to handle deployment and inferencing, which is a significant issue in AI. This could be a big opportunity for HPE, especially with smaller AI models and edge computing.
- HPE’s focus on sustainability could be a differentiator in the future. However, it faces significant hurdles such as persuading customers to move their data into an HPE environment.
Generally, we were happy that HPE avoided discussions around quantum computing at HPE Discover. Although this may surprise some, it may make sense given that quantum is not yet ready for real-world applications.
Bottom line:
HPE’s competitive edge in the AI and HPC market may lie in its infrastructure software and approach to handling large and complex models. It also has potential advantages in the deployment and inferencing aspects of AI and could stand to benefit from a future focus on sustainability. However, significant hurdles, including the need to strengthen its AI ecosystem and convincing customers to move their data, remain.
On balance, we give high marks to HPE for including LLM-as-a-service inside of GreenLake. In addition, HPE under Neri has a clear path of differentiation, which over time should pay dividends. HPE’s AI cloud offering will not be available for six months and it’s unclear how truly integrated it will be, so that is something we’ll be watching as an indicator of maturity. As well, it’s one thing to label the HPC business as AI but another thing entirely to generate profitability from the initiative.
That will be the ultimate arbiter of success.
Watch this clip of the analysts discussing the keys to watch in the future for HPE’s LLM play.
Keep in touch
Thanks to Alex Myerson and Ken Shifman on production, podcasts and media workflows for Breaking Analysis. Special thanks to Kristen Martin and Cheryl Knight, who help us keep our community informed and get the word out, and to Rob Hof, our editor in chief at SiliconANGLE.
Remember we publish each week on Wikibon and SiliconANGLE. These episodes are all available as podcasts wherever you listen.
Email david.vellante@siliconangle.com, DM @dvellante on Twitter and comment on our LinkedIn posts.
Also, check out this ETR Tutorial we created, which explains the spending methodology in more detail. Note: ETR is a separate company from Wikibon and SiliconANGLE. If you would like to cite or republish any of the company’s data, or inquire about its services, please contact ETR at legal@etr.ai.
Here’s the full video analysis:
All statements made regarding companies or securities are strictly beliefs, points of view and opinions held by SiliconANGLE Media, Enterprise Technology Research, other guests on theCUBE and guest writers. Such statements are not recommendations by these individuals to buy, sell or hold any security. The content presented does not constitute investment advice and should not be used as the basis for any investment decision. You and only you are responsible for your investment decisions.
Disclosure: Many of the companies cited in Breaking Analysis are sponsors of theCUBE and/or clients of Wikibon. None of these firms or other companies have any editorial control over or advanced viewing of what’s published in Breaking Analysis.
Photo: Sundry Photography/Adobe Stock
A message from John Furrier, co-founder of SiliconANGLE:
Your vote of support is important to us and it helps us keep the content FREE.
One click below supports our mission to provide free, deep, and relevant content.
Join our community on YouTube
Join the community that includes more than 15,000 #CubeAlumni experts, including Amazon.com CEO Andy Jassy, Dell Technologies founder and CEO Michael Dell, Intel CEO Pat Gelsinger, and many more luminaries and experts.
THANK YOU