The AI Semiconductor Landscape Primer
Michael Spencer
A.I. Writer, researcher and curator - full-time Newsletter publication manager.
To get my best coverage consider becoming a paid subscriber for less than $2 a week. To read this post with infographics and proper formatting please visit the original here.
This was written on January 20th, 2025. Happy Trump inauguration day! With the U.S. continuing a number of stringent exports controls and this next administration expected to keep it up maybe even with elevated tariffs it’s a super interesting time to think more about the semiconductor industry. The AI arms race and national security due diligence related to U.S. exceptionalism is upon us.
A flurry of Executive Orders by the Biden Administration on his last two weeks in office were telling. Trump and the new administration will be carefully watched and their actions scrutinized. Meanwhile, I’ve long admired the work of Eric Flaningam for his macro overviews on various aspects of technology stacks. Let’s feature some of them here:
?? Generative Value ??
His Newsletter, Generative Value provides great insights into how everything is connected.
Articles to check out!
Whether you are an investor, technologist or just a casual reader his overviews will provide you some value and are easy to understand and scan.
As you know, the Biden administration has implemented a series of executive orders (EOs) and export controls (ECs) aimed at regulating the semiconductor industry, particularly in response to national security concerns and competition with China. Trump has said various things with regards to tariffs on China as well. The U.S. appears to be trying to control how AI spreads to other nations, limiting China’s ability to for example access the best Nvidia AI related GPUs and chips.
Jake Sullivan — with three days left as White House national security adviser, with wide access to the world's secrets — called on journalists and news media to deliver a chilling, "catastrophic" warning for America and the incoming administration:
The AI Arms Race circa 2025
What happens this point on is fundamentally a new world of innovation and competition in innovation.
“The next few years will determine whether artificial intelligence leads to catastrophe — and whether China or America prevails in the AI arms race.”
But how does it all work? What are the companies involved? Why are companies like Nvidia, TSMC, ASML and others so pivotal? What about the big picture and landscape?
The AI Semiconductor Landscape
By Eric Flaningam, December, 2024.
Hi, my name’s Eric Flaningam, I’m the author of Generative Value, a technology-focused investment newsletter. My investment philosophy is centered around value. I believe that businesses are valued based on the value they provide to customers, the difference between that value & the value of competitors, and the ability to defend that value over time. I also believe that technology has created some of the best businesses in history and that finding those businesses will lead to strong returns over time. Generative Value is the pursuit of those businesses.
1. Introduction
Nvidia’s rise in the last 2 years will go down as one of the great case studies in technology.
Jensen envisioned accelerated computing back in 2006. As he described at a commencement speech in 2023, ”In 2007, we announced [released] CUDA GPU accelerated computing. Our aspiration was for CUDA to become a programming model that boosts applications from scientific computing and physics simulations, to image processing. Creating a new computing model is incredibly hard and rarely done in history. The CPU computing model has been the standard for 60 years, since the IBM System 360.”
For the next 15 years, Nvidia executed on that vision.
With CUDA, they created an ecosystem of developers using GPUs for machine learning. With Mellanox, they became a (the?) leader in data center networking. They then integrated all of their hardware into servers to offer vertically integrated compute-in-a-box.
When the AI craze started, Nvidia was the best-positioned company in the world to take advantage of it: a monopoly on the picks and shovels of the AI gold rush.
That led to the rise of Nvidia as one of the most successful companies ever to exist.
With that rise came competition, including from its biggest customers. Tens of billions of dollars have flowed into the ecosystem to take a share of Nvidia’s dominance.
This article will be a deep dive into that ecosystem today and what it may look like moving forward. A glimpse at how we map out the ecosystem before we dive deeper:
领英推荐
?? Generative Value Newsletter ??
A mental model for the AI semiconductor value chain. The graphic is not exhaustive of companies and segments.
2. An Intro to AI Accelerators
At a ~very~ high level, all logic semiconductors have the following pieces:
Traditionally, CPUs are general-purpose computers. They’re designed to run any calculation, including complex multi-step processes. As shown below, they have more cache, more control units, and much smaller cores (Arithmetic Logic Units or ALUs in CPUs).
On the other hand, GPUs are designed for many small calculations or parallel processing. Initially, GPUs were designed for graphics processing, which needed many small calculations to be run simultaneously to load displays. This fundamental architecture translated well to AI workloads.
Why are GPUs so good for AI?
The base unit of most AI models is the neural network, a series of layers with nodes in each layer. These neural networks represent scenarios by weighing each node to most accurately represent the data it's being trained on.
Once the model is trained, new data can be given to the model, and it can predict what the outputted data should be (inference).
This “passing through of data” requires many, many small calculations in the form of matrix multiplications [(one layer, its nodes, and weights) times (another layer, its nodes, and weights)].
This matrix multiplication is a perfect application for GPUs and their parallel processing capabilities.
The GPU today
GPUs continue to get larger, with more computing power and memory, and they are more specialized for matrix multiplication workloads.
Let’s look at Nvidia’s H100 for example. It consists of CUDA and Tensor cores (basic processors), processing clusters (collections of cores), and high-bandwidth memory. The H100’s goal is to process as many calculations as possible, with as much data flow as possible.
The goal is not just chip performance but system performance. Outside of the chip, GPUs are connected to form computing clusters, servers are designed as integrated computers, and even the data center is designed at the systems level.
Training vs Inference
To understand the AI semiconductor landscape, we have to take a step back to look at AI architectures.
Training iterates through large datasets to create a model that represents a complex scenario, and inference provides new data to that model to make a prediction.
A few key differences are particularly important with inference:
Combined, these reasons help explain why so many new semiconductor companies are focused on inference. It’s a lower barrier to entry.
Nvidia's networking and software allow it to scale to much larger, more performant, and more reliable training clusters.
On to the competitive landscape.
3. The AI Semiconductor Landscape
We can broadly look at the AI semiconductor landscape in three main buckets:
Visualizing some of those companies below:
Read the entire article here.
OK Bo?tjan Dolin?ek
Finance at Microsoft | Climate and VC | Founder at VCo2 Media
1 个月I love to see you both collaborate!
VC Investor | Actionable insights on startups, innovation, and entrepreneurship
1 个月This is a great article, congrats!
Partner at JME.vc
1 个月Interesting
Partner at Notion Capital
1 个月Jimmy Acton - check this out ??