GP Bullhound's weekly review of the latest news in public markets.
GP Bullhound
GP Bullhound is a software advisory and investment firm supported by 120 employees across 12 offices globally.
This week’s update covers key takes from the AWS re:Invent conference, more data points in AI and GPU availability, and software reporting, including Salesforce.
Our Technology Predictions 2024 report is out. Find out what's in store for the tech sector in 2024.
Market:?A generally quiet week, with some last quarter results (and year) rolling in.
Portfolio:?We significantly reduced our position in Workday this week (see comments below), though balanced by our significant portfolio position in Salesforce.?
AWS held its annual re:Invent conference this week.?Jensen Huang turned up on stage, and it all returned?to chips and Nvidia.?
The opening keynote on Monday night was about the?move to serverless computing as the next iteration in the cloud.?The way companies have run and built their business has changed significantly over the past 20 years: if you were a startup in the early 2000s, you bought Sun servers and Cisco routers and built your application on that hardware. Then cloud came along and allowed you to rent a server (and rent specific capacity) within a third-party datacentre, effectively moving up-front capex cost to ratable opex. The next step has been the move to serverless –?getting rid of servers and letting you run your workloads across a broad layer of compute and storage infrastructure, scaling automatically for the capacity you need.?It has the benefits of delivering much better efficiency and, for the end customer, being?much more cost-effective – because you only pay for the compute you use. In a world?where compute?–?driven by AI – is exploding, that becomes even more important.?Efficiency means access to cheaper compute, and more affordable AI compute will drive?more innovation and AI use cases (a la OpenAI/ChatGPT).?
Amazon already has its ASIC chips?–?Trainium and Graviton?–?and launched new generations?this week. Both new iterations feature much?more memory bandwidth?–?the Trainium2 3x more and the Graviton4 75% more than the prior generations. That goes back to the importance of memory as the performance bottleneck we’ve spoken about (and, relatedly, Micron positively pre-released this week around AI demand).?
We’ve said before that it makes sense for hyperscalers to build their own chips?– they have?enough utilisation, specific use cases, and specific workloads to apply to ASICs.?And there’s no doubt that?no player wants to be entirely tied into one very powerful supplier in Nvidia.?It’s important that they build?ASICs and not GPUs, however?– these have much higher performance but?are very specific to the application and have much more limited scope in terms of workloads that can run on them.?
While ASICs will rebalance some workloads, AI GPUs will?remain the majority of the market?because we are still?so early on in AI use cases and, therefore, the flexibility that comes with a GPU vs a custom chip is much more important. It’s not to say that there won’t be a point in the future when we have more ASIC deployments – that will likely come with more stabilisation in the applications and algorithms that run on chips (telecom base stations run on ASIC equivalents – that speaks to the maturity of that market). But, in the?short to mid-term, we expect GPUs?–?where Nvidia and AMD are really the only credible offerings?–?to represent the bulk (90%+) of AI infrastructure?and for ASICs?not to cannibalise?GPU workloads. Look at the comments below from Dell – it might not be able to get hold of any Nvidia GPUs for 39 weeks, but, equally, it can’t sub them out for ASICs.
As an aside, AMD said at a competitor conference this week that they expect?to exceed their prior $2bn target for their pure GPU offering?(we think it will come closer to $5bn).?
That gets us to?Nvidia and Jensen on stage with AWS CEO Adam Selipsky.?The reality is that while it’s more cost-effective for Amazon to run the workloads it can on its own chips,?AWS (and Google and Azure) customers want to train their models using Nvidia, as developers continue to remain tied into using CUDA.?AWS will now be the first to offer?DGX Cloud?– Nvidia’s AI training as a service – which allows enterprises to take their data and plug it into Nvidia’s pre-trained models or more easily build their models with their datasets run on the hyperscalers.?
This makes sense for Nvidia –?if more of the innovation happens on top of large language models further up the stack, the risk is that value will shift away from CUDA. Nvidia presumably hopes that customers choose to implement the full Nvidia stack, and it has both CUDA and DGX cloud as the competitive moat, ultimately controling more of the value chain.?
It’s less clear why the hyperscalers want to do this. Nvidia is trying to go direct to the customers. In a perfect world for Nvidia,?enterprises will choose Nvidia for their cloud service first and foremost. Whether running on AWS, Azure, or GCP will be secondary?– we assume that Nvidia will implement its architecture so that you won’t get wildly different experiences on whichever cloud provider it’s built upon. There is?potential for a value shift and price competition for hyperscaler capacity (why wouldn’t I go for the cheapest cloud service if Nvidia assures me the experience is the same?)
It seems like Nvidia is abstracting away the middle layer, which is the hyperscaler capacity, and capturing all the value in?both?the front end (owning the customer) and the back end (chips).?Will these partnerships continue in a more normal GPU supply world??Still, for now, Nvidia is the king master, and everyone has to play nice – the hyperscalers need as close a relationship with Nvidia as possible.?
The other announcements worth noting from the event were around LLMs and AI services,?which are particularly interesting in the context of the OpenAI debacle.?Amazon announced it is rolling out a workplace chatbot, Amazon Q, to compete directly with Microsoft’s Copilot.?The thing it’s missing is the broader integration Microsoft can offer with its full office suite, which is why it’s priced more cheaply?– $20 per month vs $30 for Copilot.?
Finally, and in the context of last week’s OpenAI news,?the focus on Amazon’s LLM service Bedrock was very much around being LLM agnostic.?“There’s not going to be one model to rule them all,” Adam Selipsky pitched.?We’ve talked about the uncertainties in LLMs – how many there will be, and will there be a winner? Will different models work better for different use cases??For Amazon, the OpenAI drama was a perfect opportunity to double down on the fact that they haven’t bet their house on one LLM (with an apparent shot fired towards Microsoft).?
Onto results and newsflow:
AI servers still GPU supply gated?–?Nvidia visibility
领英推荐
I wish I could tell you that the backlog was less than?–?or the lead time – was less than 39 weeks. I can’t today.?We are on the phone, working every available channel opportunity, as you might imagine, with our supply chain capabilities, to improve supply, to improve supply availability.?We’ve offered our services. We’ll help where we can. I’m hopeful for the day to tell you that supply has improved greatly. Lead times have reduced, and we can work the backlog down faster. That’s our job. I don’t have those answers today.?It’s 39 weeks. We’re trying to continue to get more supply. That’s where we are. As we look forward into 2024, there’s clearly alternatives coming. There’s work to be done in those alternatives, software stacks have to be taken care of, resolving the opportunities around them. But there’s more options coming. Their adoption rate, we’ll see. But right now, that?multibillion-dollar pipeline that I referenced, the backlog that we’ve talked about is Nvidia-based 39-week lead time, we’re working our behinds off every day to get more supply.
Portfolio view:?We don’t own any dedicated server providers – as per our intro on Amazon Cloud and serverless, ultimately, much more of the value either lies with the underlying chips or up the value chain at the application layer. The server providers have low barriers to entry and little pricing power. But it’s clear evidence that the AI server shift is happening,?with Nvidia chips, the key components everyone wants to get their hands on, still limited by supply.
Semis?–?AI beneficiaries beyond GPUs?
Portfolio view:?We own Marvell and see it as a clear (though perhaps less appreciated) beneficiary of AI; in some cases, its DSPs are built into AI systems at a more than one-to-one attach rate with GPUs.?
AI Memory inflexion driving pricing and revenue upgrades
Software demand holding up?–?billings and cRPO dynamics need to be monitored closely?
We significantly reduced our position in Workday in the fund this week, given a lack of conviction around short-term results and new management creating uncertainty in the outlook.?With the shares back close to their highs before downgrading its guidance in September,?we felt the risk/reward balance was to the downside. We will revisit it as we continue to reassess the risk/reward of our portfolio positions.?
More broadly in software, we’re digging around contract durations and the impact on billings/cRPO. It seems to us that there are a few?more datapoints this results season that suggest we’ve been through a period of companies securing sales with multi-year discounts,?which customers were happy to do when money was free, which made billings look good (but hurts revenues down the line) and which is now rolling off. Ultimately, rolling back to shorter duration contracts doesn’t impact the P&L and earnings. Still, it’s important to scrutinise why any metric might be deteriorating and what it might say about the underlying demand.?
Retail?–?Cyber week and China smartphone data
Portfolio view:?Still no real signs of weakness in consumer, though from a technology company perspective, it’s still an area we have limited exposure to.
For enquiries, please contact: Inge Heydorn, Partner, at [email protected] Jenny Hardy, Portfolio Manager, at [email protected] Nejla-Selma Salkovic, Analyst, at [email protected]
About GP Bullhound GP Bullhound is a leading technology advisory and investment firm, providing transaction advice and capital to the world’s best entrepreneurs and founders. Founded in 1999 in London and Menlo Park, the firm today has 14 offices spanning Europe, the US and Asia.